Comparing Several Independent Means Flashcards
What does a one-way Anova tell you?
It compares two or more independent groups. It checks whether there is an actual difference between the groups but it does not tell you which groups differ.
What can you do to see which groups differ? (2)
Contrasts tests or posthoc analysis
What assumptions does a one way anova have? (4)
- Continuous variable
- Random sample
- Normal distribution
- Homogeneity of variance (The assumptions that there is equal variance within the groups.)
How can the normality assumption and the variance assumption be tested?
The assumption of the normal distribution can be tested using the Shapiro Wilk test and the assumption of homogeneity of variance can be tested using the Leveneβs test.
What two hypotheses do one way independent ANOVAβs usually have?
h0 = There is no differences between the groups hπ = There is a difference between the groups
What test statistic does the one way independent ANOVA utilise?
F-statistic
What ratio and formula does the F-statistic utilise?
It represents a signal to noise ratio of the data and it uses the following formula:
F= MSmodel/MS error
What do the degrees of freedom of the F-statistic depend on?
The sample size and the number of groups.
What are the three types of variance in a one way independent ANOVA?
Model (between-groups)
Error (within-groups)
Total
What is the sum of squares for each of these types of variances?
πππππππ =βππ(ππ βπ)^2
πππππππ =πβ1
ππ = βπ 2π(π β1) πππππππ = N-k
πππ‘ππ‘ππ = πππππππ + πππππππ
ππ =πβ1 π‘ππ‘ππ
N denotes the sample size, ππ denotes the sample size per category and k denotes the number of categories.
What is the mean square for each of these categories?
πππππππ = πππππππ/ ππ πππππ
πππππππ = πππππππ/ ππ πππππ
πππ‘ππ‘ππ/ ππ π‘ππ‘ππ
What do the model and error variations represent in terms of explaining the variation?
The model variance represents the variance that can be explained by the experimental manipulation. The error variance represents the variance that cannot be explained by the experimental manipulation.
What are contrasts?
Planned comparisons used to investigate a specific hypothesis and compare different parts of data with each other
How are contrasts limited?
They can only use one piece of data once
Why are contrasts limited in this way?
In order to not inflate the type-I error rate
What can solve this problem?
Post-hoc tests are unplanned comparisons that compare all groups with each other and adjusts the alpha level to control for the inflated type-I error rate.
What is the difference between the ANOVA and regression between two variables?
There isnβt one
What three rules are there to develop a planned contrast?
- If you have a control group, this is there usually because you want to compare it against any other groups
- Each contrast must compare only two chunks of variation
- Once a group has been singled out in a contrast, it canβt be used in another contrast
What rules should be followed when applying weight to contrasts?
- Choose sensible contrasts
- Groups coded with positive weights will be contrasted against groups coded with negative weights
- If you add up the weights for a given contrast the result should be 0
- If a group is not involved in a contrast, automatically assign it a weight of zero, which will eliminate it from the contrast
- For a given contrast, the weights assigned to the group(s) in one chunk of variation should be equal to the number of groups in the opposite chunk of variation
What three criteria decide which post-hoc test is best?
Does the test control the type I error rate?
Does it control the type II error rate?
Is the test robust?
When should the REGWQ post-hoc test be used and when shouldnβt it?
It has good power and tight control of the type I error rate so it should be used when you want to test all pairs of means however when group sizes are different this should not be used
Comment on the type I error rate and power of the Tukey and Boneferoni test
Control the Type I error rate pretty well but lack statistical power. Boneferoni has more power when number of comparisons are small but Tukey has more power when testing large numbers of means
Are post-hoc procedures robust
Small deviances from reality they perform well, perform badly when group sizes are unequal and when population variances are different
What tests were designed for when group sizes are different? When should each of these be used?
Hochbergβs GT2 (if group sizes are very different) and Gabrielβs pairwise test (If they are slightly different)
When you have equal sample sizes and group variances are similar use ______ or ______
REGWQ; Tukey
If you want guaranteed control over type I use _______
Bonferroni
If there is any doubt that group variances are equal then use ______
Games-Howell procedure