Ch. 15 Flashcards
What is ANOVA?
Analysis of Variance, comparing multiple variances of multiple groups simultaneously
What is the null hypothesis of ANOVA?
Population means are the same for all treatments
What is the alternate hypothesis of ANOVA?
At least one mean is different from the others
What are “mean squares”?
Means squares are a special name for the measures of variation used in ANOVA
What is the group mean square?
It is a value proportional to the observed amount of variation among the group sample means.
From pg. 466, “The group mean square represents variation among the sampled individuals belonging to different groups. It will on average be similar to the erro mean square if the population means are equal.”
What is the error mean square?
Estimates the variance among subjects taht belong to the same group
What is the F ratio?
The Ratio of MSgroups to MSerror
What should F equal if the null hypothesis is true?
1 (no difference in variations)
What should F be if the null hypothesis is false?
A value greater than 1 (beause the real differences among group means will enlarge MSgroup)
Degrees of freedom for groups?
df = k - 1, where k is the number of groups
Degrees of freedom for error sum?
dferror = sum of (n-1) = N - k
where N is total number of data points in all groups combined, and k is the number of groups
(i.e. N - k is equal to the sum of the df of each individual group)
What is the test statistic for ANOVA?
F-ratio
Degrees of freedom for F-ratio?
There are 2, the numerator’s df (MSgroup df which is k-1) and the denominator’s df (MSerror df which is N-k)
thus want F(k-1), (N-k), 0.05
Note that numerator (k-1) always goes first
What is R2?
It measure the fraction of the variation in Y that is explained by group differences.
How to understand R2?
The greater the value of R2, the larger the effect of the variation in groups (variation caused by explanatory variable) on the total variable.
Can consider result as a percentage effect of treatment variation compared to total variation (pg. 469)
ANOVA w/ two groups vs. t-sample t-test
Two sample t-test and ANOVA give identical resutls. However, t-test can work with hypothesised differences between means, unlike ANOVA testing u1=u2 (ex u1 - u2 = 10). Also, Wlech’s t-test allows for
Assumptions of ANOVA?
Random Samples
Variable is normaly distributed in each population (robust, esp. w/ large sample means
Variance is the same in all populations (same 10-fold boundary as t-test)
Kruskal-Wallis test
- What is it?
- When to use it?
- Power?
Non-parametric test similar to ANOVA; based of Mann-Whitney U-test and has the same limitations/assumptions.
Similar power to ANOVA when samples are large, but little power when small. Use ANOVA if able
Probability of a Type 1 error in N tests? (equation)
1 - (1 - a)N
Bonferroni correction?
Basically tells you to use a small alpha value
a’ = a/(number of tests)
Planned comparion?
A comparison between means planned during the design of the study, identified before the data are examined (since ANOVA doesn’t tell you what means are different, just that one of the means are different, at least)
Gives more power than just a two-sample t-test
Unplanned comparison
One of multiple comparisons carried out to help determine where the differences in the means lie
Tukey-Kramer test
What is it and when is it done?
Done after finding variation among groups in single-facter ANOVA
Compares all group means w/ all other group means
Steps of Tukey-Kramer?
- Order group means in ascending magnitude
- create null hypotheses
- Perform test
Probability of making a Type-1 error in Tukey-Kramer?
The error throughout the course of testing all pairs of means is no greater than significance level alpha (that you set)
This is because a compensation for data dregging was built in
Why not do t-test w/ Bonferroni correction instead of a Tukey-Kramer?
Tukey-Kramer has more power as it uses more info from the data, so it is better than t-tests.
Also has a built in adjustment for the number of tests.
Assumptions of Tukey-Kramer?
Same as ANOVA, Random sample, all normal distriubtions, and equal variances.
However, not as robust as ANOVA since only 1 pair tested at a time.
P-value of Tukey-Kramer is exact when the design is balanced (n1=n2=n3), but conservative if not balanced (i.e actual probability of Type 1 error smaller than stated crit. value)
Fixed vs Random effects
Fixed effects have treatments chosen by the experimenter, while random effects have a random sample of treatments from all possible treatments
Single-factor ANOVA is not affected by fixed or random effects
Advantage of random-effects ANOVA?
Results are generalizable, unlike fixed-effects that can only talk about the groups included in the study .