LECTURE 3 AND 4 multiple regression Flashcards

1
Q

what is multiple regression

A

using linear regression model to predict the value on one variable from several predictor variables - hypothetical relationship between several variables

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

what equation does the multiple regression use

A

expansion of straight line equation

y = b0X1 + b2X2 ….+error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

forced entry

A

all predictors entered simultaneously

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

hierachial

A

experimenter decides input of variables based on theoretical background

aloows to observe unique predictive influence of a new variable on the outcome as known predictors held constant

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

stepwise

A

predictors selected using semi partial correlation with outcome

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

r value

A

correlation between observed value and predicted value

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

r square

A

proportion of variance accounted for by the model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

adjusted r square

A

estimate of r square in popualtion (shrinkage) - estimates the change in r square when generalise from sample to the population

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

beta values

A

chnage in outcome associated with unit change in the predictor - also has SD value (as increase value by 1SD have a ___ SD effect on second variable)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

t test in ANOVA

A

tells whether each IV make a significant contribution to predicting the DV (coeff sig diff from 0) (t= p=)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

how to interpret SD beta values

A

when predictor value increase by 1SD, predicted value increase by ___ (beta) of a SD

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

r square change (hierachial)

A

hows how the second variable has accounted for more of the models variance
ie.e r square of model 1 is how much variance accounted for by 1st vairable
r square of model 2 is how much variance accounted for by both vairable - r square change is the specific variance accounted for by the addition of the second variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

how to report anova

A

F (df, residual df)=__, p= __)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

how can you assess accuracy of the multiple regression model

A

standardised residuals and influential cases (cookes distance)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

describe standardised residuals for accuracy of the multiple regression model

A

average sample - 95% STDresiduals between +- 2
99% between +- 2.5

outliers of >3

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

describe cookes distance for accuracy of the multiple regression model

A

measures influence of a single case on the whole model

concern = absolute values > 1

17
Q

‘variable type’ regression assumption

A

outcome must be a continuous variables

predictors can be continuous or dichotomous

18
Q

‘non zero variance’ regression assumption

A

predictors cannot have 0 variance

19
Q

‘linearity’ regression assumption

A

relationship must be linear

20
Q

‘independence’ regression assumption

A

all values come from different person

21
Q

‘No Multicollinearity’ regression assumption

A

predictors must not be highly correlated
check with collinearity diagnostics - values should be:
TOLERANCE = > 0.2
VIF =

22
Q

‘Homoscedasticity’ regression assumption

A

for each value of predictor - variance of error should be constant
check by plot of ZRESID against ZPRED to get normality of errors probability plot

23
Q

‘independent errors’ regression assumption

A

for any observation, error terms should be uncorrelated

24
Q

‘normality distributed errors’ regression assumption

A

histogram and normal probability plot of standardised residuals