Multiple Regression Flashcards

1
Q

Properties of fitted regression line

A
  1. exogeneity assumption
  2. no-correlation between residuals and fitted values
  3. mean of residuals = 0
  4. variance decomposition = explained variance + unexplained variance
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
1
Q

Assumptions

A
  1. Normality of residuals - histogram, plots of residuals, formal tests
  2. Homoscedasticity - homogeneity of variance of residuals - scatterplots of residual variance, Levene’s test
  3. Linearity - scatterplot
  4. Non-correlated errors
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Cohen’s f values

A

0.02 - small
0.15 - medium
0.35 - large effect

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Measures of model fit

A
  1. R2 - might overestimate the explained variance while we add other variables
  2. adjusted R2 - can decrease
  3. R2 change
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Types of Influential Observations

A
  1. outliers
  2. leverage points
  3. influential observations - reinforcing, conflicting, shifting
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Multicollinearity - impact

A
  1. Estimation - unique variance of each predictor is reduced. when they’re not correlated - total explained variance = sum of unique explained variances by each predicot
  2. Explanation - problem in interpretation

sometimes small multicollinearity can be ignored

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Handling multicollinearity

A
  1. delete
  2. combine
  3. estimation techniques
  4. do nothing but don’t intepret in a regular way
  5. increase sample size
  6. dimensionality reduction
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Suppression

A
  1. inflated R^2 - very high
  2. high correlation between predictors, but only one of them is correlated with dependent
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Correlations

A
  1. Zero order - unique + shared -> (b+c)/total
  2. Partial - x&y but controlled for other variables -> b/total - (a+c)
  3. semi partial - unique variance of x -> b/total -> R^2 will decrease by this amount if we remove this variable from the model
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

R^2 will decrease by which correlation?

A

squared part/ semi-partial - unique contribution of the variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

suppression & correlation

A

when partial is higher that zero-order, we have suppression

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

how can including suppressor improve the model?

A
  1. it can remove the irrelevant variance in the predictor - decrease the noise
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Interpretation with interaction

A
  1. conditional main effects
    e.g. the income for females is less than males under the condition that years of education is zero
  2. interaction term
    females earn 60 euros multiplied by years of education less compared to men
  3. difference in slope = slope of … is different
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

interaction and follow-up analysis

A
  1. interaction with metric - simple slope
  2. interaction between metric and non-metric - simple slope
  3. interaction between non-metrics - simple effect
How well did you know this?
1
Not at all
2
3
4
5
Perfectly