SPECIFICATION BIAS Flashcards

1
Q

Parsimony

A

Occam’s razor suggests that model be kept as simple as possible because a model can never truly capture reality.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What are the attributes of a good model

A
  1. Parsimony
  2. Identifiability
  3. Goodness of Fit
  4. Theoretical Consistency
  5. Predictive Power
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Identifiability

A

Only one estimate per parameter for a given set of data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Goodness of Fit

A

Explanatory variables explain as much of the variation in Y as possible; least squared residuals, meaning a high as possible adjusted R-squared

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Predictive Power

A

The model whose theoretical predictions are borne out by experience.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Theoretical Consistency

A

When constructing a model it should have some theoretical underpinning. The coefficients must then reflect theory and have the correct sign.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Specification Errors

A
  1. Omission of relevant variables
  2. Unnecessary variables
  3. Wrong Functional form
  4. Errors of measurement
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Omission of Relevant Variables

A
  1. X2 & X3 are correlated, then a2 is biased { E(a2)= B2 + B3b32 (where b32 is the interaction). Ea1 = B1 + B3 (Xbar3- b32Xbar2). Unless b32 is zero, then the estimator is biased.
  2. a1 & a2 are inconsistent no matter the sample size
  3. If X2 & X3 are uncorrelated, a2 unbiased & consistent. (Estimators can be consistent and not unbiased).
  4. Variance for the estimator will not be the same: E[vara2]= var (b2) + B23 sum x2/ (n-2) sum x2
  5. Even when uncorrelated, the variance will be biased and overestimate the true variance; this means a positive bias and wider confidence interval–more likely to accept the null even when shouldn’t.
  6. Intercept will be biased and underestimate the true intercept. Also, the Standard Error will be incorrect.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Upward bias and Downward bias

A

bias or deviation from the true parameter, positive results in an upward bias, and negative results in downward bias.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Why is Omission of Relevant Variables bad?

A

It overestimates the impact of a variable , and prevents the model from determining the true impact of the explanatory variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Inclusion of Irrelevant variables

A

Overfitting; including unnecessary variables will retain the unbiased and consistent nature, however the variance and in turn the SE will be larger, expanding the confidence interval; less efficient and precise–potential to not reject the null. *may lead to multicollinearity e.g. X’s that do the same thing.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Incorrect Functional Form

A

Log-linear model; A2 measures the elasticity of Y with respect to X2. Whereas the regular linear model tells you the slope or rate of change; unclear what model it is suggesting we use* follow-up

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Errors of Measurement

A

Nonresponse errors, computing errors, reporting errors;
If the error is in the dependent variable: bs (OLS estimators) are unbiased, variance unbiased, but variance is larger because u incorporates the error in the dependent variable.
Error in the Explanatory Variable: OLS estimators are biased; inconsistent.
Error in both: very serious
solutions: instrumental or proxy variables

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Consistent

A

As the the sample gets larger, the estimator becomes unbiased

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Instrumental or Proxy Variables

A

Substitutes for the variable itself when it is difficult to collect accurate values; these variables are highly correlated with X variables, uncorrelated with the measurement error & u.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What causes specification errors?

A
  1. poorly formulated model
  2. weak theory
  3. Don’t have the necessary data.
17
Q

Testing a Point estimate

A

Use the T distribution and use the critical t and calculated t-value to determine whether to reject the null hypothesis.

18
Q

Testing a Joint estimate (two variables)

A

Use the F distribution to determine if the variables are relevant. e.g. B3=B4=0 (null hypothesis) Do an F test.

19
Q

Test for Omitted Variables or Incorrect Functional Forms

A

Same as others situations: examine adjusted R, estimated t ratios, and coefficients (are they in line with theory?)

20
Q

Examination of Residuals (test)

A

Plot the residuals (e) of the fitted model : should be random and tight (remember no autocorrelation); if there is a pattern, typically means there is an additional variable not included that can explain some of the variation in Y.

21
Q

Ramsey’s RESET (regression error specification) Test

A

Helps to identify if the model is misspecified, but it doesn’t help figure out what the alternate model should be.

  1. Obtain the estimated Y, e.g. Yhat.
  2. Insert Y back into the equation: Yt = B1 + B2Xt+ B3Yhat2t +B4Yhat3t + vt
  3. regress this model to get the Rnew
  4. Use Ftest to determine significance: Rnew-Rold/new regressors divided by 1-Rsquared-new/ n-# of parameters)
22
Q

Ramsey’s RESET (regression error specification) Test

A

Helps to identify if the model is mispecified, but it doesn’t help figure out what the alternate model should be.

  1. Obtain the estimated Y, e.g. Yhat.
  2. Insert Y back into the equation: Yt = B1 + B2Xt+ B3Yhat2t +B4Yhat3t + vt
  3. regress this model to get the Rnew
  4. Use Ftest to determine significance: Rnew-Rold/new regressors divided by 1-Rsquared-new/ n-# of parameters)