Chapter 12: General Linear Model, Comparing Several Different Means Flashcards
What does an ANOVA measure?
analysis of variance
when you are interested in comparing differences between the means in 3 or more independent groups
are population means the same or different?
want to know whats generally true about the groups, not too interested in the means of the sample (mostly mu)
Why wouldn’t we use multiple independent samples t-tests instead?
increased chance of type 1 error (declaring there is an effect when there really is not)
familywise alpha increases over .05
FW alpha is the conditional probability of making one or more type 1 errors when performing multiple tests
Are an ANOVA and regressions separate things?
No. Regression is a more general form of an ANOVA. Anything you can do with an ANOVA you can do with a regression.
Regression can handle both categorical and continuous predictors, while ANOVA can only handle categorical
the idea they are different is partially historical (regression = applied research; ANOVA = experimental)
What are the methods for comparing independent means?
to keep FW <.05/maintaining solid power
the more liberal the alpha, the greater the power
- Confidence intervals
- standard linear model w/dummy coding (comparing groups to a base condition)
- One way ANOVA
- Welch or Brown-Forsythe F (Fbf)
- Planned Contrasts
Confidence Intervals
Positives:
-simple/straightforward
-makes you think about magnitude/focuses on estimation and avoids b&w thinking
Negatives:
- risk finding a difference when one doesn’t exist (type 1 error), increased with sample size
- with a small number of groups, the CIs are not going to be as sensitive to differences that do exist
- not as sensitive/powerful as other analyses
- if CIs intersect too much, you can’t really draw conclusions
- error bars represent an estimate of where the population mean is
- larger bars = more uncertainty in the data. smaller number of cases means bigger error bars
interpretation: we are 95% confident that the population mean for X is between A & B
Standard linear model with dummy coding
comparing groups to base condition
involve the use of a regression model
useful when you want to compare groups to base/control group
use k-1 to make a base group (number of groups -1)
check the magnitude of R^2 change and whether that change is significant
negatives:
- compare groups only with the control group
- since they are both comparing against some standard (control), they are not independent tests
interest in mu difference, if so, how dif
R^2 change tells us group membership accounts for about X% of the variance in IV
a sif f change means the means are significantly different from eachother
In standard linear model with dummy coding on SPSS…
Start by looking at the f test. if it is significant, then look at the regression coefficients to see where those differences are. if it is not significant, you do not have to look at the coefficients.
allows you to see which means are different
Logic of F statistic
- tests overall fit of a linear model to a set of data
- when model is based on group means, our predictions from the model are group means
- different group means? good prediction, F will be high
- similar group means? not a good prediction, F will be low (~1), fail to reject null
compare improvement in fit due to using the model from the grand mean
if the differences between group means are large enough, then the resulting model will be a better fit to the data than the grand mean (null)
Linear Model: Overview/Summary
- model of “no effect” or “no relationship between the predicot and outcome” is one where the predicted value of the outcome is always the grand mean
- we can fit a different model to the data that represents an alternative hypothesis. we compare the fit of this model to the fit of the null (i.e., using the grand mean)
- the intercept and one or more parameters (b) describe the model
- the parameters determine the shape of the model thay we have fitted. therefore, the bigger the coefficients, the greater the deviation between the model and the null model (grand mean)
Group means: Overview/Summary
- in experimental research the parameters (b) represent the differences between group means
- if the differences between group means are large enough, then the resulting model will be a better fit to the data than the null model (grand mean)
- if this is the case: predicting scores from group membership better than simply using the grand mean. in other words, the group means are not all the same
calculating an f statistic
- quantify the amount of variability in the scores
- ## separate variability into 2 parts: the part that can be accounted for by group membership and the part that cannot be accounted for by group membership
more people = more residual
more groups = bigger model score
SStotal equation
total amount of variability in the scores
SSt = s^2grandmean(N-1)
s^2grand = (Xi-Xgrand)^2
square each difference then add them all together
SSmodel equation
how much variability accounted for by the model/group membership
SSm = n(Xgroupmean-Xgrandmean)^2
do for each group and add them all together
SSresidual equation
how much variability isn’t accounted for by the model/group membership
SSr = s^2group(n-1)
or
SSr = SSt - SSm
do for each group then add them together
n is number of people in each group
MSm equation
remove the effect of the # of groups/people
important because the SS get bigger the more groups/people you have
MSm = SSm/DFm
DFm = number of groups -1
k-1