Lecture 2 Flashcards

1
Q

MLR difference to SLR

A

Started adding new explanatory variables to improve accuracy of model and reduce bias in estimating the effect of the primary variable of interest

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What do the coefficients of each explanatory variable mean and how do we get them?

A

They represent the marginal effect of each corresponding variable, get them by finding the partial derivative with respect to corresponding variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

In order to give a causal interpretation to the coefficients the key assumption is:

A

The expected value of u given all values of the xis is 0
- i.e. when you control for these, there’s no systematic bias in the error term

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

SST =
R^2 =

A

SST = SSE + SSR

R^2 = SSE/SST = 1 - SSR/SST

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

R^2 value in relation to explaining the variation in the dependent variable

A
  • r^2 can never decrease when adding more regressions, as assign another explanatory variable can only either improve or leave unchanged the amount of variability explained by the model
    CAREFUL - don’t overfit, as focusing on maximising R^2 may lead to the inclusion of irrelevant variables just to improve the model’s fit artificially.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

MLR.1

A

The model is linear in parameters

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

MLR.2

A

We have a random sample of size n from the population

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

MLR.3

A

No perfect collinearity, so in the sample, none of the explanatory variables are constant, and there are no exact linear relationships between them (perfect collinearity)
- without this OLS would not be able to distinguish the independent contribution of each variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

MLR.4

A

Zero conditional mean
- if u is correlated with the explanatory variables, it is said to be endogenous, making biased and inconsistent estimates.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Steps to analyse OLS in MLR:

A
  1. Obtain convenient representations
  2. Manipulate expression to obtain Bj^ = Bj + sampling error
  3. Compute mean and variance of sampling error

Use the partialling out method

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Partialling out algorithm for computing OLS estimators

A
  1. Regress x1 on x2
  2. Compute residuals
  3. Regress the outcome y on the residuals obtained above

The estimator found in step 3 above is numerically identically to the MLR estimator, we are essentially isolating the pure effect of x1 on y, holding x2 constant

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Omitted Variable Bias

A

Why it’s a problem can be shown by trying to find expected value of B1, when u includes another explanatory variable within it
- will make the estimator biased

= B1 + B2.01
Where 01 = cov^(x1,x2)/Var^(x1)
- so if B2 = 0, or the covariance is 0, the SLR will get it right

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

MLR.5

A

Homoskedasticity, so the variance of the error, u, doesn’t change with xi

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Sample variance of OLS slope estimators under MLR

A

Var(BJ^|Xn) = o^2/SSTj(1-Rj^2)
- lower error variance means estimator more precise
- higher variation in Xj, means estimator more precise
- lower R^2, means less xj is correlated with each other, easier to isolate its effect on y - leading to a smaller variance for estimator

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Unbiased o^2 estimation under the GM assumptions

A

O^^2 = (1/n-k-1). (Sum of ui^s squared) = SSR/df

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Standard error is computed as

A

Se(Bj^) = o^/(SSTj(1-R^2))^0.5

17
Q

Why OLS always preferred to the alternative under GM?

A

It is both unbiased and efficient, and MLR.5 means the variance is lower than the alternative, making it a more reliable estimator.

BLUE - minimising the sum of squared residuals, while utilising the assumption of homoskedasticity to achieve the most precise estimates.