Week Nine - Multiple Linear Regression Flashcards

1
Q

What are the 5 assumptions of linear regression?

A

CORRECT VARIABLES: interval
INDEPENDENCE OF DATA: each person should only participate once
SAMPLE SIZE/NORMALITY: larger SS the better
LINEARITY: can produce zero correlations if not linear
HOMOSCEDACITY OF RESIDUALS

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is homoscedacity of residuals?

A

The error variance should be the same at each level of the predictor

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

How can we test for normality?

A

QQ plots - should be a straight line

normality test under assumption checks on J

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

With residual plots (under assumption checks) what are we hoping the data to look like?

A

Rectangle with no pattern or outliers

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

When we have heteroscedacity imply? and what do the tests need to be?

A

That the model is only good for certain scores

p values need to be >.05

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

How do we test for outliers? (2)

A

Basic approach: any residual value > 3 SD from mean

Cooks distance: A measure of the influence of one case on the model as a whole
values > 1 may be a concern

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is colinearity?

A

When some of the IVs are closely related meaning that they provide little unique information

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What does the model coefficients box tell us in simultaneous multiple regression?

A

They assess whether each regression coefficient is significantly different from zero in the context of other predictors - p>0.05 not useful

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is tolerance?

A

How much of the variability in the predictor variable is not explained by the other predictors (values <0.1 are a problem - most of the variation in the IV is explained by the others)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What VIF scores are a problem?

A

values >10

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the underlying component of hierarchical regression?

A

Whether a predictor can add to the prediction of an outcome variable beyond the amount that is already explained by a particular predictor

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What does the output of a HMR tell us?

A

it will create a change in R2 scores which will tell us how much the variation in the out come is explained by adding the extra predictor

Will also give delta f score explaining how much extra variation in %, and a p value

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What does HMR examine? and what is it driven by

A

The incremental importance of a predictor variable

It is driven by hypotheses - researcher will specify

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What do you use when you don’t have a hypothesis for your regression?

A

Use a step-wise approach

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is a forward stepwise entry approach?

A

Where we put the best predictors into the model first and then only entering more predictors if they improve the quality of the predictive model (ie increases R2 significantly)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is a backward stepwise approach?

A

Where we start out with all the predictors in the model and then start throwing out the worst ones until this has a negative impact on the quality of the predictive model (ie reducing R2 significantly)