Multiple Regression Flashcards
Prediction
Not an experiment
Survey data
Relationships between things
Built on correlation analysis
In regressions always have at least 1 predictor (simple linear) or more (multiple)
Variable terms
Criterion=y
Analogous to DV (how labelled in SPSS)
Predictor(s)= x
Analogous to IV
One predictor
Y=mx+c (y=bx+a) gives line of best fit
Multiple predictors
Y=b1X1+bX+a
Y=b2X2+bX1+bX+a
(B1X1)= predictor 1
(BX)= predictor 2
b shoes strength of relationship
Regression
Equation/formula/model
In correlation/regression, looking at association between outcome measure & 1 predictor
In multiple regression, looking at association between criterion & set of predictors
Model is full set of predictors & how you put them together to try & predict criterion (outcome measure)
The Prediction
R= correlation strength (-1-1 with 1 being perfect positive relationship)
The same as ‘r’ in simple correlation
Adjusted R square= variance explained
r/r^2= correlation or single regressions R/R^2= multiple regression
Covariance
Covariance= how jointly variable 2 variables are
Also affected by SD
Tells you how much overlapping variance there is between 2 factors
SD’s tell you how much variance there is for each individual factor
Divide covariance by SD’s to get r/R
r^2/R^2
r/R x r/R
Gives you proportion of variance in the criterion variable that is explained by regression
Adjusted R^2
Takes account of sampling bias (how big sample is, offsets data of v small sample)
Controls for the effects of sample size
Tells you what proportion of the variance in criterion variable you can predict (explain) with your predictor or model
Use of regression
Good for things that can’t be measured experimentally, e.g. hard to measure or unethical
Stats test (ANOVA) used to see if prediction is better than chance Reported like normal ANOVA, if significant then better than chance
Beta
B= strength between x & y value Beta= normalised/standardised version of B, shows effect size, as Beta gets bigger t gets bigger (t does same thing as beta but less important)
Predictors need to be standardised so can be compared to one another, beta does this
B’s can go below -1 & above 1 but beta’s can’t
Beta shows most useful predictors (explained the most variance in the predictor)
How to write it up
A standard multiple regression was conducted with {DV} as the criterion variable & {IV’s} as the predictor variables
The model accounted for {R^2}% (adjusted) of the variance
A significant model emerged (F(a,b)=c, p=d).
Significant contributions were made by _____ (predictor _), beta=a, t=b, p=c, and ___ (predictor _), beta=a, t=b, p=c.
General notes
Only report p values to 3 dp, everything else to 2 dp
Italicise all stats letters in Latin alphabet
Don’t italicise Greek letters e.g. beta
Assumptions
1) more predictors= more participants (atleast 15 pp per predictor)
2) do predictors correlate with the DV?
3) Do IV’s correlate too much (>0.8) with each other?- multicollinearity,
4) do you have unequal spread of error?- heteroscedasticity (similar residuals, similar to homogeneity if variance)
Additivity & linearity
The outcome variable is linearly related to any predictors
If you have several predictors then their combined effect is described by adding their effects together
If this assumption is not met then your model is invalid