Chapter 11 part a: GLM1 Flashcards
1
Q
ANOVA
A
- Linear model to compare several means
2
Q
Predictor with 2 categories
- b
A
- represents the difference between the mean of 2 categories
- is the difference statistically different?
3
Q
Predictor with more than 2 categories
- b
A
- we have to create dummy variables so that b compares differences between two means
- each dummy variable will have 2 categories
4
Q
ANOVA vs Regression
A
- we use ANOVA to test fit of regression line
- ANOVA: special case of linear model (regression)
- Equation: Same
Outcome=model+error
5
Q
Important in ANOVA:
A
- baseline category sample size must be fairly large so b retains reliability
6
Q
ANOVA Equation
A
Outcome = bo + b1Dummy1 +b2Dummy2
- bo: mean of base category
- b1: difference between control mean and 1st group to compare to mean
- b2: difference between control mean and 2nd group to compare to mean
7
Q
ANOVA: F significant
A
- using group means to predict is better than using overall mean
8
Q
Logic of F-ratio
A
- if group means are the same: our model is poor [F small]
- if group means are different: our model is good [F large]
- F: whether group means are different
9
Q
Logic of ANOVA
A
- Simplest Model: Grand Mean of outcome: No effect
- Intercepts and parameters describe the model
- Parameters: shape of fitted Model
—> the bigger the coefficients (b): the greater the deviations between model & grand mean - Experimental research: b represents the difference between group means
- If the difference between groups is large enough then our model is better than grand mean
10
Q
Total Sum of Squares [SST]
A
- total amount of variation
- SST=sum(obs. data-grand mean)^2
11
Q
Grand Variance
A
- variances between all scores regardless of experimental condition
- grand variance (s^2)=SST/(N-1)
- SST=grand variance x (N-1)
12
Q
Model Sum of Squares [SSM]
A
- variation explained by the regression model
- SSM=sum[nk(group mean-grand mean)^2]
- dfM= k-1
13
Q
Residual Sum of Squares [SSR]
A
- variation that can NOT be explained by our model
- variation caused by extraneous factors
—> SSR= SST-SSM —> SSR=sum(xk- group mean)^2 —> SSR= SSR1+ SSR2+ SSR3... —> SSR=sum[sk^2 (nk-1)] - variance of each group x (nk-1) —> dfR= dfT-dfM —> dfR= N-k
14
Q
dfT
A
N-1
15
Q
dfM
A
k-1
16
Q
dfR
A
N-k
17
Q
ANOVA
A
- Omnibus test: whether explained variance is larger than unexplained variance
- significant F: means of categories are not equal (several scenarios possible)
—> manipulation has had some effect but what the effect is: Does NOT say
18
Q
Mean Squares
- why do we need them?
A
- SS are biased
- MS eliminate their biases
19
Q
Model Mean Square
A
- SSM/dfM
- systematic variation
20
Q
Residual Mean of Squares
A
- MSR=SSR/dfR
- unsystematic variation
21
Q
F-ratio
A
- how well the model is against how bad it is
- F= MSM/MSR
- F>1: if model good but check its significance
22
Q
Assumptions of One-WAy ANOVA
A
- Linearity
- Normality
- Homoscedasticity: check using Levene’s and correct using Brown or Welch
~ F(bf): SSR=sum[sk^2 (1-nk/N)]
~ both techniques control for Type 1
~ Welch: best except when very large variance - Independence of Errors
23
Q
Is ANOVA robust?
A
- controls somewhat for Type 1
- controls for skew, kurtosis, non-normality for 2tailed
- less control for 1 tail
- leptokurtic: type 1 error too low
- platykurtic: type 2 too high
24
Q
ANOVA Robust to Normality?
A
- Yes, when group sizes are equal
- When group sizes are unequal: F and power may be affected
25
ANOVA robust to heteroscedasticity?
- Yes, when sample sizes are equal
| - No, when sample sizes are unequal