week 8.Analysis of Variance Flashcards
ANOVA
(Analysis of Variance) allows researchers to test the significance of the differences between 2 or more means.
Xij=µ + rj + €ij
where i and j are groups, r=treatment effect for group and €=error.
ie Basic concept of predicting a value, given group mean, grand group mean, and error.
Assumptions of Anova
- Assume variances of each group are the same (if 2 populations have the same variance, we assume sample variances will be alike)
- Assumes that the variables are normally distributed
- assumes that the scores of one group are entirely independent of scores from another group.Also, each individual score within a group should be entirely independent of any other particiapnts in group.
Logic of Anova
- If assume the variances of groups are the same, and assume normally distributed, then groups will only differ in their mean.
- Total variability in the data has 2 components:
a) variability associated with the treatment
b) error variance
Error Variance
Provides an index of variability between people who have been treated the same. ie Within one group, there will be chance variations of motivation, personality, IQ, and experimental error etc. Also called Within-Group Variation.
Between-treatments Variance
Also called Treatment variance.Differences between the means of the groups may be due to the different treatment, or chance.
F-statistic
For ANOVA, we use and F-statistic. We need to determine if the differences between treatment groups are larger than would be expected by chance alone.
F=variance between treatments/error variance
F=treatment effect+differences due to chance/differences due to chance
ie, if there is no treatment effect (H0 is true), then the F ratio will equal 1. If there is a treatment effect, F>1.
F=MStreatment/MSerror
ANOVA calculations
SS=Sum of squares=Σ(X-X-)2
For Anova, we have SStreatment
SSerror and
SStotal .
SStreatment=nΣ(X-j -X-…)2
where n=group sample size,
X-j=group mean
and X-…=grand mean.
SSerror=Σ(Xij-X-j)2
where Xij=score of person i in group j
and X-jis the mean of group j.
SStotal=SStreatment + SSerror
Then if we divide an SS by it’s respective degrees of freedom, we get Mean Squares (MS).
SS/df=MS
dftotal=N-1=dftreatment + dferror
dftreatment=k-1 where k=number of treatments
dferror=k(n-1)
Evaluating the significance of the F-statistic
Use the Fcritical tables.
the degrees of freedom for treatment are the numerator degrees of freedom
and the error degrees of freedom are the denominator degrees of freedom.
If F(dftreatment,dferror) >Fcritical, then we reject the null hypothesis.
Effect size
To evaluate the effect size of the Anova finding, we use partial Eta squared (n2). This can be interpreted as a squared correlation coefficient. eg if n2=0.23, then 23% of variation in dependent variable can be attributed to differences in the variable which was manipulated across different groups.
In general. n2=0.01=small effect
n2=0.06=medium effect
n2=0.14=large effect.
Comparison between Anova and t-test
Anova allows comparisons between >2 group means. For comparison of group A, B and C,
would need 3 seperate t-tests (A vs B, Avs C, and Bvs C) whereas can do 1 Anova. If do multiple t-tests, increase the liklihood of a type 1 error, simply because each single test has a 5% chance of error.
Levene’s test
Levene’s test us is to test the assumption that different groups are of homogenous variance. (one of the assumptions of ANOVA.)
If the results of Levene’s (also denoted by F) has an accompanying p value and p < 0.05 then it is concluded that the assumption of homogenity has been violated. Some might then recommend using different statistics to anova.