W7 multiple regression analysis Flashcards
Ordinary least squares regression
When we have > 1 predictor (X) variable, we simply need a way to combine them
The regression procedure expands seamlessly to do this…
Essentially, we extend the least squares procedure to estimate b0 and b1 and b2… bk to give us the best possible prediction of Y from all the variables jointly
This type of regression might be referred to in some texts as OLS, Ordinary Least Squares Regression
Equation for multiple regression
Y’ = b0 + b1X1 + b2X2 + e
Testing the significance of R
fancy p = rho
-> population correlation coefficient
H0 : fancy p = 0
H1 : fancy p not equal to 0
H0 : there is no linear relationship between X and Y in the population
H1: there is a linear relationship between X and Y in the population
R squared and variance
In correlation, r squared = proportion of variance shared by X and Y, i.e., the two correlates—overlap
Multiple R squared and variance
- Multiple R2 is variance in Y accounted for by Xk
usually referred to simply as R2 - Not specific to any one single X, it is for the set of them.
- The regression equation is also referred to as the Model
we check the value of R2 in the Model Summary box in the SPSS output
R squared equation
R squared = SSreg / SStot
- these can be manually calculated (i guess)
- found in ANOVA table
SStot
Y - Y bar
SSreg
Y’ - Y bar
SSres
Y - Y’
Extending our interpretation of regression predictors
Semipartial correlations—sr and sr2
- > Relationship between Xk and Y when all other predictors partialled out of only Xk.
- > Good indicator of the unique relationship between Xk and Y
- > Always squared (i.e., sr2) to give proportion of unique variance in Y accounted for by Xk
Calculating the unique variance
- unique variance of X1 on Y is equal to X1’s “part” squared, found in correlations coefficients table when using zpp command
- unique variance of X2 on Y is equal to X2’s “part” squared from correlations table
- the “part” has removed the variance accounted for from the other variable and gives the variance (when squared) of that variable alone
Shared variance
- the difference between the combined ‘unique variance’ (sr squared X1 + sr squared X2) and ‘total variance explained’ (R squared)
Shared variance = R squared - (sr squared X1 + sr squared X2)