3. MLR - Estimation Flashcards

1
Q

What are the advantages of a multiple regression over a simple regression?

A
  • Incorporate more explanatory factors into the model
  • Explicitly hold fixed other factors that otherwise would be in the error term u
  • Allow for more flexible functional forms
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What are slope parameters?

A

Parameters other than the intercept

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

How do the methods used change for a regression that has a quadratic function?

A

Mechanically there will be no difference in using the mehtod of OLS to estimate the parameters but there is a difference in the interpretation of the coefficients

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

How can a model be more flexible?

A

A model can contain logs and quadratic forms suggesting different relationships between the explanatory variable and the dependent variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is the partial effect?

A

Where you fix one variable and see what happens to y when the other variable changes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What do residuals measure?

A

The difference between the true value and the fitted value

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is Frisch-Waugh theorem?

A

1) Regress the explanatory variable x on all other explanatory variables and obtain the residuals r^
2) Regress the dependent variable y on the resulas r^ to obtain B^

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Why does this procedure work?

A
  • The residuals from the first regression is the part of the explanatory variable that is uncorrelated with the other explanatory variables
    The slope coefficient of the second regression therefore represents the isolated effect of the explanatory variable
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What are the standard assumptions for the multiple regression model?

A

MLR.1 - Linear in parameters
MLR.2 - Random sampling
MLR.3 - No Perfect collinearity
MLR.4 - Zero conditional mean

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is perfect collinearity?

A

If an independent variable is an exact linear combination of the other independent variables then we say the model suffers from perfect collinearity and it cannot be estimated by OLS. Some exact relationship between the regressors.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is key to remember about the MLR.3 assumption - no perfect collinearity?

A

Assumption MLR.3 does allow the independent variables to be correlated; they just cannot be perfectly correlated

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Why, in a multiple regression model is the likelihood of the assumption of zero conditional mean more likely to hold?

A

Because fewer things end up in the error term u

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What does over-specifying the model mean?

A

One (or more) independent variable is included in the model even though it has no partial effect on y in the population

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Why do we use OLS estimates?

A

Because they are good enough on average that they will be equal to the true value

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

How do OLS estimates relate to population parameters?

A

The OLS estimators are unbiased estimators of the population parameters

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

How do you interpret unbiasedness?

A

The estimated coefficients may be smaller or larger depending on the sample that as a result of it being random however on average they will be equal to the values that characterise the true relationship

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

What is the issue regarding overspecifying the model?

A

Including irrelevant variables may increase sampling variance thus reducing the precision of the estimates but it has no effect on the level of unbiasedness

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Alongside poor specification of our model, when else can MLR.3 Perfect collinearity fail?

A

Assumption MLR.3 also fails if the sample size, n, is too small in relation to the number of parameters being estimated.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

What is underspecifying the model and why is it problematic?

A

Excluding a relevant variable or underspecifying the model causes the OLS to be biased

20
Q

Between simple and multiple regression analysis, which one is more likely to have omitted variable bias?

A

Simple

21
Q

What are EXOGENOUS explanatory vairables?

A

When assumption MLR.4 (Zero conditional mean) holds

22
Q

What are ENDOGENOUS explanatory variables?

A

If xj is correlated with u for any reason, then xj is said to be an endogenous explanatory variable. The term “endogenous explanatory variable” has evolved to cover any case in which an explanatory variable may be correlated with the error term.

23
Q

What is the difference between MLR.3 and MLR.4?

A

Assumption MLR.3 rules out certain relationships among the independent or explanatory variables and has nothing to do with the error, u. You will know immediately when carrying out OLS estimation whether or not Assumption MLR.3 holds. On the other hand, Assumption MLR.4—the much more important of the two—restricts the relationship between the unobserved factors in u and the explanatory variables. Unfortunately, we will never know for sure whether the average value of the unobserved factors is unrelated to the explanatory variables.

24
Q

What notation change do we observe when we know we are underspecifying our model?

A

We use the symbol “~” rather than “^” to emphasise that ߘ1 comes from an underspecified model.

25
Q

In the context of omitting a variable from your analysis, when do we say a sample has upward bias?

A

In the context of omitting a variable if E(ߘ1) > ß1, then we say that ߘ1 has an upward bias

26
Q

In the context of omitting a variable from your analysis, when do we say a sample has downward bias?

A

When E(ߘ1) < ß1, ߘ1 has a downward bias. These definitions are the same whether ß1 is positive or negative.

27
Q

In the context of omitting a variable from your analysis, when do we say a sample is biased towards zero?

A

The phrase biased toward zero refers to cases where E(ߘ1) is closer to zero than is ß1.

28
Q

What type of bias do we identify if ß1 is positive and biased towards 0?

A

If ß1 is positive, then ߘ1 is biased toward zero if it has a downward bias.

29
Q

What type of bias do we identify if ß1 is negative and biased towards 0?

A

On the other hand, if ß1 < 0, then ߘ1 is biased toward zero if it has an upward bias.

30
Q

What does correlation between a single explanatory variable and the error usually suggest?

A

Correlation between a single explanatory variable and the error generally results in all OLS estimators being biased. The only exception to this is when x1 and x2 are also uncorrelated.

31
Q

What is homoskedasticity?

A

The error u has the same variance given any value of the explanatory variables. In other words, Var (u|x1, …, xk) = σ^2.

32
Q

For a given dependent variable, what is the only way to reduce error variance?

A

For a given dependent variable y, there is really only one way to reduce the error variance, and that is to add more explanatory variables to the equation (take some factors out of the error term). Unfortunately, it is not always possible to find additional legitimate factors that affect y.

33
Q

What is multicollinearity?

A

High (but not perfect) correlation between two or more independent variables is called multicollinearity

34
Q

What is the systematic part of a model?

A

The part consisting of explanatory variables

35
Q

What is imperfect multicollinearity?

A

The sampling variance of the slope estimator for xj will be higher when xj can be better explained by the other independent variables

36
Q

What does omitted variable bias result in?

A

You will end up being very precise to something that isn’t true.

37
Q

What happens to the slope estimator under perfect multicollinearity?

A

The variance of the slope estimator will approach infinity

38
Q

How can multicollinearity be reduced?

A

Grouping similar variables or by dropping some independent variables however this may lead to omitted variable bias

39
Q

What is the choice to include a specific variable motivated by?

A

The choice can be made by analysing the tradeoff between bias and variance

40
Q

What does OLS are BLUE mean?

A

Best Linear Unbiased Estimator

41
Q

What assumptions have to be satisfied for OLS to be BLUE?

A

The Gauss-Markov assumptions

42
Q

How do you find which unbiased estimator has the smallest variance?

A

In order to answer this question one usually limits oneself to linear estimators, i.e estimators linear in the dependent variable

43
Q

What will be the best prediction of y?

A

It’s conditional expectation

44
Q

What happens in the model if a relevant variable is excluded from the model?

A

This induces a bias in B^~1 unless x1 and x2 are uncorrelated

45
Q

What is the cost of including an irrelevant variable in the model?

A

A higher variance for the estimator of B1

46
Q

How do you calculate the degrees of freedom?

A

Count the number of parameters, including the intercept, and subtract this amount from the number of observations. (In the rare case that an intercept is not estimated, the number of parameters decreases by one.)

47
Q

How can we detect multicollinearity?

A

Using variance inflation factors (vif)