week 4 Simple ANOVA Flashcards
Power
The probability of making a type 2 error (failing to reject the null hypothesis when it is in fact false) is called beta. NOT the same as beta weight etc. The Power is the sensitivity of an experiment. The Power is the ability to detect a significant effect, in our data.
Power = 1-beta.
A powerful/sensitive experiment is one with a low probability of making a type 2 error.
Normally want Power to =0.8.
Factors which will increase Power:
- increase alpha level (eg from 0.05 to 0.1).NOT advisable in practice.
- Increasing number of subjects per condition (Anova) or overall (regression).
- A large effect size.
- Controlling extraneous influences (background noise).
Anova assumptions
The assumptions of Anova may not always be met, but the more the assumptions are met, the more confidence we have that the analysis can be run and interpreted as intended.
- Assumes variances of each group are roughly the same. The largest variance from any group SHOULD NOT be greater than 4 x the smallest variance from any group. So eg Anova can cope with the largest variance in a group being 400 and the smallest variance in a group being 100.
- Assumes continuous variables are normally distributed
- All scores are independent of all others.
Also note that in Anova, the predictor values are categorical, whereas in multiple regression, they may be continuous or categorical or mixture.
Anova is used when we have one DV and one IV with more than 2 levels. If we had one DV and an IV with 2 levels, we would run an independent samples t-test.
Anova asks if there is a significant difference in group means, but does NOT tell us WHERE the difference is.
basic model
Is SStotal which compares observed DV scores against the Grand Mean. All the variance around the Grand mean.
best model
is SSM which compares group mean against the Grand mean.
SSmodel=SSbetween
Note that calculating SSmodel is slightly different if have unequal group sizes.
residuals
Is SSresiduals which compares participant’s dv score to the group mean. All the variance within the group, around the group mean.
SSresiduals=SSwithin
independence of errors
If observations are truly independent (as is desired), then each possible error is also independent.
homogeneity of variance
Levene’s test us is to test the assumption that different groups are of homogenous variance. (one of the assumptions of ANOVA.)
If the results of Levene’s (also denoted by F) has an accompanying p value and p < 0.05 then it is concluded that the assumption of homogenity has been violated. Some might then recommend using different statistics to anova.
Occasionally Levene’s is too sensitive, so can run a Hartley’s Fmax ratio to see if really have violated.(if largest variance not 4x as large as smallest variance, is still ok). If conclude variance has been violated, can correct F statistic with either Welch Correction (corrects degrees of freedom calculated for each group;use this if have uneven group sizes) or Brown-Forsythe Correction (which recalculates SS by doing around median as opposed to mean).
non parametric test
a test which has no assumptions about data distribution (ie no assumption of normal distribution)
magnitude of experimental effect
a very large F value, does NOT necessarily mean a very big result.
With a large enough sample size, even the most trivial effects will show a very large f value.
The magnitude of the difference between 2 groups, in standard deviation units, is known as the effect size, and this tells us how big (the magnitude) of the effect. the larger the effect size, the greater the magnitude of the finding.
Cohen’s d, partisl eta squared and omega squared all measure the effect size.
Bear in mind that any effect size also needs to be interpreted in relation to the previous researchin any particular topic area. The interpretation of significance of “large” and “small” effect size can vary widely with specific situation.
Monte Carlo studies
Computer-generated studies designed to violate Anova’s assumptions to some degree and thus test the robustness of Anova.
violations of assumptions
In actuality, violations of the Anova assumptions are fairly common.
Moderate violations of skew (skew is assessment of if left and right balanced=non skew)can be handled by Anova but rate of type 1 errors is affected by kurtosis.(kurtosis = level of heavy or light-tailed around normal distribution)
unequal group sizes causes problems, particularly as the discrepancy increases.
Type 1 errors also increase if there is non independence of scores.
Welch procedure
An adjustment made when homogeneity of variance is violated.aims to reduce type 1 error.
high variability
in high variability situations, even though groups have a different mean, as there may be high overlap, it is hard to tell with any score, which group it most likely belongs to. In this situation, the mean is not a great indicator of the group. Therefore, knowing the mean difference is not enough, we also need to know how indicative the mean is for the group.
Kruskal-Wallis non parametric test
a test used to rank data irrespective of group, then tests if ranks differ with group.
Calculations
SS=Sum of squares=Σ(X-X-)2
For Anova, we have SStreatment
SSerror and
SStotal .
SStreatment=nΣ(X-j -X-…)2
where n=group sample size,
X-j=group mean
and X-…=grand mean.
SSerror=Σ(Xij-X-j)2
where Xij=score of person i in group j
and X-jis the mean of group j.
SStotal=SStreatment + SSerror
Then if we divide an SS by it’s respective degrees of freedom, we get Mean Squares (MS).MSresidual is an estimated measure of the population variance. MSmodel is an estimation of the population variance when the null hypothesis is true.
SS/df=MS
dftotal=N-1=dftreatment + dferror
dftreatment=k-1 where k=number of treatments
dferror=k(n-1)