More Regression Flashcards

1
Q

What is required to make predictions using a regression model?

A

Gauss-Markov assumptions should be met

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is a dummy variable?

A

A variable which summarises a category with discrete options e.g. D could be a dummy variable for sex with male corresponding to D = 0 and female to D = 1, in which case

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

How could regression be used for a model where there is a mean male wage and a mean female wage?

A

If wm = αm + εm and likewise for females, a dummy variable set to 0 for males would make it so that w = Dαf + (1 - D)αm + Dεf + (1 - D)εm = αm + D(αf - αm) + εm + D(εf - εm) which can be written α + Dβ + u with the intercept being the male wage and the slope being the difference in means

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is functional form?

A

The method to go beyond a linear model so that logs can be used with CRM

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is given by the slope in the relationships Y = α + βX + ε, ln(Y) = α + βX + ε, and ln(Y) = α + ln(β)X + ε?

A

The (constant) average change to Y from each unit of X
The % change in Y associated with each additional unit of X regardless of the initial level (xi + 1 –> yi + 100β) based on the approximation for a small change in a log, called the semi-elasticity of Y wrt X
The constant % change to Y from a 1% increase in x, called the constant elasticity of Y wrt X (so β = 1 means 1% increase in X increases Y by 1%)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

When is a log transformation used?

A

When it turns a non-linear relationship to a linear relationship

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

How can the total sum of squares of a regression model be broken down?

A

Σ(Y - Ȳ)2 = Σ(Y - Ŷ)2 + Σ(Ŷ - Ȳ)2
Total sum of squares TSS = Error sum of squares ESS + Regression sum of squares RSS
To prove this, use Y = a + bX + e and Ȳ = a + bX̄ to get an expression for TSS which will include Σ2b(X - X̄)e which is assumed to equal zero for OLS
Using b(X - X̄) = Ŷ - Ȳ leaves the required relation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is R2?

A

R2 = RSS/TSS = 1 - ESS/TSS gives the % variation in Y that is explained by the model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is multiple regression?

A

When there is more than one random variable on the RHS of the regression (multiple regressors)
A systematic difference in residuals of two groups after bivariate regression would suggest there should be another regressor

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is the non-parametric method for predicting Y conditional on X1 and X2?

A

Calculate a mean wage for every possible paired value of Xs and use that as the CEF E(Y|X1, X2) for distinct X pairs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the parametric method for multiple regression?

A

Use the OLS condition to get FOCs for Ŷ = a + bX1 + cX2
∂SSR/∂a = -2Σe = 0
∂SSR/∂b = -2ΣeX1 = 0
∂SSR/∂c = -2ΣeX2 = 0
Substituting e = Y - Ŷ gives a = Ȳ - bX̄1 - cX̄2

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

If using a dummy variable for X2 where 1 is true, what can be said about the CEF?

A

E(Y|X1, X2 = 0) = a + bX1 and E(Y|X1, X2 = 1) = a + bX1 + c so c is the difference between the latter and the former (the intercept of the CEF shifts but the slope doesn’t change)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is the Frisch Waugh Lovell Theorem?

A

A method used to estimate residuals when trying to estimate a model with two regressors but only interested in one
Regress Y on X1 to leave residuals eY and regress X2 on X1 to get residuals ex, then regress Y residuals on X residuals and the coefficient on the latter will be the same as c from the multiple regression

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

How does the Frisch Waugh Lovell Theorem work?

A

Regressing Y on X1 and X2 on X1 partials out the affects of other independent variables (X1), since residuals in a linear regression represent the left-over variation in the dependent variable after accounting for one independent variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly