ANOVA 1 - University of South Florida

Download Report

Transcript ANOVA 1 - University of South Florida

One-Way ANOVA
Introduction to Analysis of Variance
(ANOVA)
What is ANOVA?



ANOVA is short for ANalysis Of VAriance
Used with 3 or more groups to test for MEAN
DIFFS.
E.g., caffeine study with 3 groups:






No caffeine
Mild dose
Jolt group
Level is value, kind or amount of IV
Treatment Group is people who get specific
treatment or level of IV
Treatment Effect is size of difference in means
Rationale for ANOVA (1)




We have at least 3 means to test, e.g., H0: 1 = 2
= 3.
Could take them 2 at a time, but really want to test
all 3 (or more) at once.
Instead of using a mean difference, we can use the
variance of the group means about the grand mean
over all groups.
Logic is just the same as for the t-test. Compare the
observed variance among means (observed
difference in means in the t-test) to what we would
expect to get by chance.
Rationale for ANOVA (2)
Suppose we drew 3 samples from the same population.
Our results might look like this:
Three S amples from the S ame P opulation
Note that the
means from the 3
groups are not
exactly the same,
but they are
close, so the
variance among
means will be
small.
4
M ean 1
3
M ean 3
M ean 2
2
1
St andar d Dev G r oup 3
0
-- 20
20
-- 10
10
00
Raw Scor es ( X)
10
Rationale for ANOVA (3)
Suppose we sample people from 3 different populations.
Our results might look like this:
Note that the
sample means are
far away from one
another, so the
variance among
means will be
large.
Three S amples from 3 D iffferent P opulations
4
M ean 1
3
M ean 2
M ean 3
2
S D Group 1
1
0
- 20
- 10
0
Raw S cores (X )
10
20
Rationale for ANOVA (4)
Suppose we complete a study and find the following
results (either graph). How would we know or decide
whether there is a real effect or not?
Three S amples from 3 D iffferent P opulations
Three S amples from the S ame P opulation
4
4
M ean 1
3
M ean 1
3
M ean 2
M ean 3
M ean 3
M ean 2
2
2
1
1
S D Group 1
St andar d Dev G r oup 3
0
0
-- 20
20
-- 10
10
00
Raw Scor es ( X)
10
- 20
- 10
0
10
20
Raw S cores (X )
To decide, we can compare our observed variance in
means to what we would expect to get on the basis
of chance given no true difference in means.
Review


When would we use a t-test versus 1-way
ANOVA?
In ANOVA, what happens to the variance in
means (between cells) if the treatment effect
is large?
Rationale for ANOVA
We can break the total variance in a study into
meaningful pieces that correspond to treatment effects
and error. That’s why we call this Analysis of Variance.
Definitions of Terms Used in ANOVA:
XG
The Grand Mean, taken over all observations.
XA
The mean of any level of a treatment.
X A1
The mean of a specific level (1 in this case)
of a treatment.
Xi
The observation or raw data for the ith person.
The ANOVA Model
A treatment effect is the difference between the overall,
grand mean, and the mean of a cell (treatment level).
IV Effect  X A  X G
Error is the difference between a score and a cell
(treatment level) mean.
Error  X i  X A
The ANOVA Model:
Xi  XG  ( X A  XG )  ( Xi  X A )
An individual’s
A treatment
The
grand
is
+ or IV effect +
score
mean
Error
The ANOVA Model
Xi  XG  ( X A  XG )  ( Xi  X A )
The grand
mean
A treatment
or IV effect
Error
ANOVA Data by Treatment Level
30
Frequency
The graph shows the
terms in the equation.
There are three cells or
levels in this study.
The IV effect and error
for the highest scoring
cell is shown.
40
20
10
0
Grand Mean
Error
IV Effect
Treatment Mean
ANOVA Calculations
Sums of squares (squared deviations from the mean)
tell the story of variance. The simple ANOVA designs
have 3 sums of squares.
SStot   ( X i  X G )
2
SSW  ( X i  X A )2
The total sum of squares comes from the
distance of all the scores from the grand
mean. This is the total; it’s all you have.
The within-group or within-cell sum of
squares comes from the distance of the
observations to the cell means. This
indicates error.
SSB   N A ( X A  X G )2 The between-cells or between-groups
SSTOT  SSB  SSW
sum of squares tells of the distance of
the cell means from the grand mean.
This indicates IV effects.
Computational Example: Caffeine on
Test Scores
G1: Control
G2: Mild
G3: Jolt
Test Scores
75=79-4
80=84-4
70=74-4
77=79-2
82=84-2
72=74-2
79=79+0
84=84+0
74=74+0
81=79+2
86=84+2
76=74+2
83=79+4
88=84+4
78=74+4
Means
79
84
74
SDs (N-1)
3.16
3.16
3.16
Xi
Total
Sum of
Squares
XG
( X i  X G )2
G1
75
79
16
Control
77
79
4
M=79
79
79
0
SD=3.16
81
79
4
83
79
16
G2
80
79
1
M=84
82
79
9
SD=3.16
84
79
25
86
79
49
88
79
81
G3
70
79
81
M=74
72
79
49
SD=3.16
74
79
25
76
79
9
78
79
1
SStot   ( X i  X G )
Sum
2
370
In the total sum of squares, we are finding the
squared distance from the Grand Mean. If we took
the average, we would have a variance.
SStot   ( X i  X G )
2
Relative Frequency
0.5
0.4
0.3
0.1
Grand Mean
0.0
Low
H igh
Sc ores on the D ependent Variable by Group
Xi
Within
Sum of
Squares
XA
( X i  X A )2
G1
75
79
16
Control
77
79
4
M=79
79
79
0
SD=3.16
81
79
4
83
79
16
G2
80
84
16
M=84
82
84
4
SD=3.16
84
84
0
86
84
4
88
84
16
G3
70
74
16
M=74
72
74
4
SD=3.16
74
74
0
76
74
4
78
74
16
SSW  ( X i  X A )2
Sum
120
Within sum of squares refers to the variance within
cells. That is, the difference between scores and their
cell means. SSW estimates error.
SSW  ( X i  X A )2
Relative Frequency
0.5
0.4
0.3
0.1
C ell or Treatment Mean
0.0
Low
H igh
Sc ores on the D ependent Variable by Group
XG
XA
Between
Sum of
Squares
( X A  X G )2
G1
79
79
0
Control
79
79
0
M=79
79
79
0
SD=3.16
79
79
0
79
79
0
G2
84
79
25
M=84
84
79
25
SD=3.16
84
79
25
84
79
25
84
79
25
G3
74
79
25
M=74
74
79
25
SD=3.16
74
79
25
74
79
25
74
79
25
SSB   N A ( X A  X G )2
Sum
250
The between sum of squares relates the Cell Means to
the Grand Mean. This is related to the variance of the
means.
SSB   N A ( X A  X G )
2
Grand Mean
Relative Frequency
0.5
0.4
0.3
0.1
C ell Mean
C ell Mean
C ell Mean
0.0
Low
H igh
Sc ores on the D ependent Variable by Group
ANOVA Source Table (1)
Source
SS
Between
Groups
250 k-1=2
Within
Groups
Total
df
MS
SS/df
250/2=
125
=MSB
120 N-k=
120/12 =
15-3=12 10 =
MSW
370 N-1=14
F
F=
MSB/MSW
= 125/10
=12.5
ANOVA Source Table (2)



df – Degrees of freedom. Divide the sum of
squares by degrees of freedom to get
MS, Mean Squares, which are population
variance estimates.
F is the ratio of two mean squares. F is
another distribution like z and t. There are
tables of F used for significance testing.
The F Distribution
F Table – Critical Values
Numerator df: dfB
dfW
1
2
3
4
5
5 5%
1%
6.61
16.3
5.79
13.3
5.41
12.1
5.19
11.4
5.05
11.0
10 5%
1%
4.96
10.0
4.10
7.56
3.71
6.55
3.48
5.99
3.33
5.64
12 5%
1%
4.75
9.33
3.89
6.94
3.49
5.95
3.26
5.41
3.11
5.06
14 5%
1%
4.60
8.86
3.74
6.51
3.34
5.56
3.11
5.04
2.96
4.70
Review




What are critical values of a statistics (e.g.,
critical values of F)?
What are degrees of freedom?
What are mean squares?
What does MSW tell us?
Review 6 Steps
1.
2.
Set alpha (.05).
State Null &
Alternative
1  2  3
H0:
H1: not all  are =.
3.
Calculate test statistic:
F=12.5
4.
5.
6.
Determine critical value
F.05(2,12) = 3.89
Decision rule: If test
statistic > critical value,
reject H0.
Decision: Test is
significant (12.5>3.89).
Means in population are
different.
Post Hoc Tests




If the t-test is significant, you have a
difference in population means.
If the F-test is significant, you have a
difference in population means. But you
don’t know where.
With 3 means, could be A=B>C or A>B>C or
A>B=C.
We need a test to tell which means are
different. Lots available, we will use 1.
Tukey HSD (1)
Use with equal sample size per cell.
HSD means honestly significant difference.
 is the Type I error rate (.05).
MS W
HSD  q
NA
Is a value from a table of the studentized range
statistic based on alpha, dfW (12 in our example)
and k, the number of groups (3 in our example).
q
MSW
NA
Is the mean square within groups (10).
Is the number of people in each group (5).
10
HSD.05  3.77
 5.33
5
From table
MSW
Result for our example.
NA
Tukey HSD (2)
To see which means are significantly different, we
compare the observed differences among our means to
the critical value of the Tukey test.
The differences are:
1-2 is 79-84 = -5 (say 5 to be positive).
1-3 is 79-74 = 5
2-3 is 84-74 = 10. Because 10 is larger than 5.33, this result
is significant (2 is different than 3). The other differences
are not significant. Review 6 steps.
Review


What is a post hoc test? What is its use?
Describe the HSD test. What does HSD
stand for?
Test
Another name for mean square is
_________.

1.
2.
3.
4.
standard deviation
sum of squares
treatment level
variance
Test
When do we use post hoc tests?

a. after a significant overall F test

b. after a nonsignificant overall F test

c. in place of an overall F test

d. when we want to determine the impact of
different factors