Econometrics 1 Flashcards

0
Q
  1. Example of a spurious relationship
A
  1. sunspots is related to business cycles

2. snowfall in Portland and annual growth rate in U.S. investments.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
1
Q
  1. Linear regression: Y = α0 + α1χ + ε

Describe causation

A

Causation is assumed to flow from the RHS to the LHS. Variations in χ systematically cause variations in Y.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q
  1. Can econometrics show causation?
A

No. It can only show correlation. A logical and well-motivated mechanism is needed to CAUSALLY link the explanatory variables and the dependent variable. Without a “well motivated story,” a regression will be spurious.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q
  1. Define “identification.”
A

Identification is knowing that something is what you say it is. The estimate of a parameter is an estimate of that parameter and not, in fact, something else.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q
  1. Give an example of the identification problem and what can cause it.
A

Effect of school vouchers on educational outcomes. People who are more likely to do well in school signed up for vouchers & all that signed up received a voucher. The estimator was not identified because voucher effects were conflated with motivational and capability effects. Cause: “comparative advantage selection” or “self-selection bias.”

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q
  1. What are the five elements that must be done correctly to achieve “good” empirical analysis.
A
  1. Specification: Choosing RHS variables and function form in logical and thoughtful manner (having a well-motivated story).
  2. Identification: Ensuring that empirical effects causally associated with each variable are well-defined (avoiding conflation).
  3. Data set: Obtaining data that is best structured for analysis (type of variation, sample size…)
  4. Estimator: Choosing best estimation rule for the model and data structure.
  5. Testing and validation: Providing info that demonstrates strengths and drawbacks of empirical model and analyais.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q
  1. In a regression analysis, the error term ε captures what?

[Kennedy, p. 3]

A

The error or disturbance term is the stochastic/random part of the model. The error term is justified in three main ways:

  1. Omission of influence of innumerable chance events
  2. Measurement error of included explanatory variables
  3. Human indeterminacy.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q
  1. What are the six assumptions of the linear regression model (Gauss-Markov Theorem assumptions)?
A
  1. Linear in the parameters and the error term
  2. Full rank: Explanatory variables are not perfectly correlated
  3. Exogeneity of explanatory variables: No X is correlated with the error term.
  4. Homoskedasticity and non-autocorrelation
  5. Data in X may be any mixture of constants and random variables, BUT must be generated by a mechanism that is unrelated to ε. [Data generation]
  6. Normal distribution: disturbances are normally distributed.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q
  1. On what does precision of the estimates depend?
A
  1. The amount of variation in Y and X contained in a sample (dependent and explanatory variables) –> More variation in the raw data- the better.
  2. Sample size (more is better).
  3. “Noise” level in process… which is related directly to variance of the error term.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q
  1. Define “precision”
A

Degree to which repeated measurements under unchanged conditions show the same results. Precision has to do w/ how compressed (small) the sample distribution is for OLS estimates. Precision declines as multicollinearity increases (error variance will increase). Precision increases w/sample size. Precision will decrease as the overall fit of the model improves (better fit results in less error therefore error variance will be reduced).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q
  1. What is a consistent estimator?
A

An estimator whose distribution converges on the true β as sample size increases. (Relates to an asymptotic estimator.) Consistency is large-sample (asymptotic distribution) counterpart to unbiasedness in finite, small-sample distributions.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q
  1. What is an efficient estimator?
A

An estimator that has the lowest variance of all other estimators in its class. Typically, where one unbiased estimator can be found- many other unbiased estimators are also possible. It is therefore desirable to chose the unbiased estimator that has the least amount of variance- aka “BEST unbiased” estimator. A researcher would be more confident that a single draw out of a distribution w/ less variance was closer to the true β than a single draw out of distribution w/large variance. [Kennedy, p. 16]

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q
  1. Why are estimators restricted to be a linear function of the observations on the dependent variable?
A

Reduce task of finding the efficient (smallest variance) estimator to mathematically manageable proportions. [Kenndy, p. 17]

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q
  1. Define BLUE (as an estimator that is BLUE).
A
B = "best" = efficient--> least amount of variance amongst all other estimators in the same class.  Requires GM assumptions 1-4 hold for OLS LUE.
L = Linear
U = unbiased: β-hat = β
E = estimator
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q
  1. Define unbiased.
A

The mean of the sampling distribution of the estimator is equal to the true parameter value.

“On average” an estimator will correctly estimate the parameter in question; it will not systematically under- or over-estimate the parameter (Greene, p. 54).

Unbiasedness is a property of finite (small-sample) least squares.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q
  1. What are “residuals”?
A

ε = y - y(hat). Observed value - predicted value.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q
  1. Define OLS
A

Ordinary least squares: The estimator generating the set of values of the parameters that minimizes the sum of squared residuals.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q
  1. Why is OLS popular?
A

NOT necessarily because it makes residuals “small” by minimizing the sum of squared errors, but because (1) it scores well on other criteria such as R^2, unbiasedness, efficiency, mean square error and (2) computational ease.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q
  1. Define and describe R^2.
A

R^2 is the coefficient of determination. It represents the amount of variation in the DV “explained” by variation in the independent variables.
R^2 = SSR /SST -or- R^2 = 1- (SSE/SST)
CAVEAT: Because OLS minimizes sum of squared residuals, it also automatically maximizes the R^2 value.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q
  1. Define “SST”
A

Total sum of squares: Σ(y - ybar)^2
y = observed value
ybar = mean of observed values

SST = SSR + SSE

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q
  1. Define “SSR”
A

Sum of squares due to regression (explained variation).
SSR = Σ(yhat - ybar)^2
yhat = estimated values
ybar = mean of estimated values

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q
  1. Define “SSE”
A

Sum of squared errors/residuals (unexplained variation): Σ(y - yhat)^2.
y = observed values
yhat = estimated values

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q
  1. What are three common data problems?
A

(1) Outliers. Can drop outliers to check difference in results, i.e. how much impact outliers had on estimation. If outliers can be explained, can augment model to take into account causes of outliers.
(2) Missing data. Don’t make up data. Make sure no self-selection resulting in bias, e.g. unemployment stats that don’t account for people who’ve given up looking for employment.
(3) Multicollinearity. Examine W.M.S. Run pre-estimate. Do hypothesis testing and drop variables with low impact on estimation. (**Start w/ more variables and drop variables w/low impact. Don’t start w/fewer and add….). If can’t break multicollinearity- GIVE UP or if have resources do own experiment.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q
  1. Is an OLS estimator a random variable?
A

Yes- because estimator (βhat) will vary each time the experiment is run because of the error term ε.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q
  1. Under what data generating conditions, does the Gauss-Markov theorem/assumptions apply and then do asymptotic properties apply along with the Grenander conditions?
A

GM properties used with finite, small-sample –> Non-stochastic X or experimental types of data.

Asymptotic properties used w/ stochastic X or non-experimental data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Q
  1. Interpret slope parameter on a standard linear model:

weight (kg) = -114.3 + 106.5*Height (meters)

A

For every 1 meter increase in height, weight will increase by an average of 106.5 kg.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
26
Q
  1. Interpret a log-log or log-linear regression model.
    log(STARTS) = C(1) + [-0.2869 * log(MR)]
    What is this type of model called?
A

Estimates correspond to point elasticities.

A 1% change in MR will cause housing starts to decrease by 0.2869%.

27
Q
  1. Interpret slope parameter on a semi-log linear regression.
    1) ln(weight) = 2.14 + 0.00055 * height
    2) Weight = 3.94 + 1.16 * ln(height)
A

1) Log-Lin Model: A one-unit (e.g. meter) increase in height results in a 0.055% increase in weight. (multiply parameter times 100)
2) Lin-Log Model: A 1% increase in height results in 0.0116 increase in weight. (divide parameter by 100).

28
Q
  1. Interpret the t-statistic in an OLS regression.
A

T-stat calculated by assuming the null hypothesis is true.
t = (b - β)/SE-hat, where b = estimated value, β is true value (assumed to be zero under the null), SE-hat is the estimated standard error of b.

The t-stat is used to determine the probability that the null is true.

If b = 0.44; SE-hat = 0.29; t-stat = 1.5; and p = 0.13  ...
THEN b (0.44) is t-stat (1.5) standard error units from β. "Think of b as a distance unit from β."  One tail test: divide p value in half (eviews): 0.13 /2 = 0.065.  If null is true, have a 6.5% chance of getting a t-ratio that is 1.5 or larger.  6.5% of all samples will have a true null (β = 0), so if reject null, have a 6.5% chance that you INCORRECTLY rejected null.

t-stat can only test one restriction at a time.

29
Q
  1. Hypothesis testing. Describe p[Type I error], confidence level, power, and P[Type II error].
A

p[Type I error] = α –> reject null, but null is true.
–> state of the world: Incorrectly reject the null when it is true.

Confidence Level = 1 - α
–> state of the world: correctly accept the null when it is true.

p[Type II error] = β –> fail to reject null, but null is false.
–> state of the world: Incorrectly reject alternative hypothesis.

Power of the test = 1 - β
–> state of the world: correctly accept the alternative when it’s true.

30
Q
  1. What may cause large standard errors?
A
  1. Multicollinearity and low variation in raw data. Large standard errors can result in insignificant parameter estimates.
31
Q
  1. What is the F-statistic and when is it used?
A

F-Stat = [(SSE(r) - SSE(u)) / J] / [(SSE(u)) / (T - K) ~ F(J(T-K))
–> Prob[F(J, (T-K)) > F-Critical] = p-value.
If F(critical) < the F-Stat, then REJECT the null.
**F-Stat is always a right-tailed test.

SSE(r) = sum of squared errors from restricted model
SSE(u) = sum of squared errors from unrestricted model
J = # of restrictions (# of = signs in the null hypothesis)
T = sample size
K = # of parameters including the slope.

F-stat = t-stat^2.

F-stat is used with finite-sample distributions. It can test more than one restriction at a time (overall model fit).

32
Q
  1. What is the Wald statistic and when is it used?
A

Wald-Stat = [SSE(R) - SSE(U)] / [SSE(U) / (T-K)] ~ Χ^2(J)

χ^2 = F-stat * J, where J is the number of restrictions.

Applies to distributions with large-sample properties.
Wald-stat is also a measure of overall model fit; it can test more than one restriction at a time. Wald is a little more relaxed than F-Stat (or T-Stat).

33
Q
  1. What is the Gauss-Markov assumption 1 and what are implications of violation?
A

GM-1: Linearity in the parameters and error.
Violations: (1) Inclusion of irrelevant variables (the mean square error can increase) or omission of relevant variables (biased estimators).
(2) Non-linear specification (Estimates are biased and without meaning).
(3) Changing parameter values. e.g. parameter values may change over time in time series or panel data; parameter values may differ by geographic or political region; included variables may be influenced by variables outside of the model.

34
Q
  1. What is the Gauss-Markov assumption 2 and what are the implications of violation?
A

GM-2: Full rank: Explanatory variables are not perfectly correlated (& have more observations than independent variables), i.e. no exact linear relationship between explanatory variables.
Violation: Exact multicollinearity (two or more Xs perfectly correlated). The OLS process breaks down mathematically (analogous to a divide by zero error) –> may get message: “near singular matrix” (a singular matrix cannot be inverted and OLS requires inversion).

An approximate linear relationship between explanatory variables results in multicollinearity (but not perfect multicollinearity) which can lead to estimation problems. Multicollinearity arises out of the particular data being used in the sample rather than theoretical or actual linear relationship among any of the regressors. Can arise due to: variables share a common time trend, EV may be the lagged value of another EV that follows a trend, some EVs may vary together ‘cuz data not collected from a wide enough base, or there could be some type of approximate relationship among regressors.

Implications: OLS estimator in presence of multicollinearity is still BLUE. R^2 unaffected. Major problem is that the variance of OLS estimates of the collinear variables is very large, thus precision is very low –> not enough independent variation in a variable to calculate w/confidence the effect it has on the DV.

35
Q
  1. What is the Gauss-Markov assumption 3 and what are implications of violation?
A

GM-3: Exogeneity of explanatory variables –> No X is correlated with the error term. Expected value of the error term is zero or the covariance between the error term and each explanatory variables is zero.

Endogeneity caused by comparative advantage selection, omitted explanatory variables (IF correlated w/included variable), reverse causation, simultaneity (eg. supply/demand), wrong functional form.

Violations will cause the OLS estimate of the intercept to be biased, parameters are not identified (cannot be estimated w/out conflation).
Asymptotic sample- OLS estimator is inconsistent.

36
Q
  1. What is the Gauss-Markov assumption 4 and what are implications of violation?
A

GM-4: Homoskedasticity and non-autocorrelation.
Disturbance terms all have the same variance (a constant) and are not correlated with one another. “Spherical disturbances”

Violations: Heteroskedasticity –> Disturbances do not have the same variance. Autocorrelated errors –> Disturbances are correlated w/one another.
Consequences of heteroskedasticity: Wrong inferences due to biased standard errors (interval estimation and hypothesis testing cannot be trusted), loss of efficiency (estimates are no longer BLUE, OLS doesn’t provide estimate w/the smallest variance).

Consequences of autocorrelation: OLS estimators are still unbiased and linear, but they no longer have the minimum variance property; usual formulas to estimate variances are biased (can have - or + autocorrelation); confidence intervals and hypothesis tests based on the t and F distributions are unreliable; R^2 is affected; computed variances and standard errors of forecasts may be inaccurate.

37
Q
  1. What are common causes of heteroskedasticity?
A

Heteroskedasticity can be caused by violation of other GM assumptions, but assuming other GM assumptions are met the following can be causes of heteroskedasticity:

Causes of heteroskedasticity: (1) high value of explanatory variable is a necessary, but not sufficient condition for a change in the dependent variable, e.g. family income and vacation expenditures; (2) values of EV become more extreme in either direction; measurement errors (some respondents providing more accurate responses than others); subpopulation differences or other interaction effects (e.g. effect of income on expenditures for different races; differences in expenditures may be smaller for low income, but get larger as income increases); model mis-specification.

38
Q
  1. What are common causes of autocorrelation?
A

(1) Inertia is data series (e.g. long-term inertia in people being unemployed); model specification errors; cobweb phenomenon (e.g. implementation of supply decisions take time to implement so agricultural commodities react to price w/ a lag of one time period; data manipulation (interpolation or extrapolation of data); nonstationarity (e.g. mean, variance, covariance change over time).

39
Q
  1. What is the Gauss-Markov assumption 5 and what are implications of violation?
A

GM-5: Stochastic or nonstochastic data: Data may be any combination of constants or random variables.

Implications for violation: ?

40
Q
  1. What is the Gauss-Markov assumption 6 and what are implications of violation?
A

GM-6: Normal distribution –> Disturbances are normally distributed (zero mean and constant variance).

Normality assumption is reasonable in most settings because of Central Limit Theorem. Useful implication of GM-6 is that it implies observations on ε(i) are statistically independent as well as uncorrelated. Normality is useful in constructing confidence intervals and test stats.

41
Q
  1. What is the choice of the best estimator of β based on?
A

Statistical properties of the candidates (different estimators) such as unbiasedness, consistency, efficiency, and their sampling distributions. (Greene, p. 51).

42
Q
  1. Describe finite-sample properties.
A
  1. Finite-sample properties of the least squares estimator are independent of the sample size.
  2. Unbiasedness is a finite-sample property.
  3. Linear model is one of relatively few settings in which definite statements can be made about the exact finite-sample properties.
  • -> E[b] = β –> Estimator is unbiased.
  • -> E[s^2] = σ^2 –> Disturbance variance estimator is unbiased.
  • -> Var[b|X] = σ^2(X’X)^-1 and Var[b] σ^2E[(X’X)^-1]
  • -> Gauss-Markov Theorem: The MVLUE of w’β is w’b for any vector of constants, w.
43
Q
  1. Describe large-sample (asymptotic properties) of least squares estimator.
A

Consistency –> Improves upon the concept of unbiasedness in two ways: (1) Except for least squares linear regression model (MVUE), it is rare for an estimator to be unbiased; (2) Property of unbiasedness does not imply that more info is better then less in terms of estimation of parameters.

Grenander Conditions for Well-Behaved Data apply to large-sample properties.

44
Q
  1. When do the Grenander Conditions for well-behaved data apply and what are they?
A

Apply for large-sample (asymptotic) properties of least squares estimator. Assumptions are very weak and likely to be satisfied by any data set encountered in practice.

G1: Sums of squares will continue to grow as the sample size increases. No variable will degenerate to a sequence of zeros.
G2: No single observation will ever dominate x’x, and as n –> infinity, individual observations will become less important.
G3: Full rank condition will always be met.

45
Q
  1. In OLS regression what how do the standard error and variance relate?
A

The standard error is the square root of the variance: sqrt(σ^2).
The estimated variance, s^2, is the square of the standard error.

The standard error is the square root of the main diagonal elements of the OLS variance-covariance matrix.

46
Q
  1. List six benefits of large samples
A
  1. More likely to get a consistent estimator (asymptotic properties)
  2. Increased precision /reduced variance
  3. Increased power
  4. Can break or reduce multicollinearity
  5. Increase in normality: Improve hypothesis testing.
  6. Reduce influence of outliers: law of large numbers.
47
Q
  1. What is e’e?
A

Sum of squared residuals.

48
Q
  1. With hypothesis testing (OLS) the restricted model is the model under the null or alternative hypothesis?
A

Restricted model is the model under the null hypothesis.

49
Q
  1. What tests can be used to test model specification?
A
  1. Davidson-MacKinnon non-nested test (Homework # 6)

2. Ramsey Reset Test

50
Q
  1. Describe Davidson-MacKinnon non-nested test for functional form.
A
  1. Start w/ two alternative models (e.g. model 1 has ln(x) and model 2 is linear (x)).
  2. Assume one model is true (here Model 2- linear x).
  3. Estimate Model 1 [ln(x)] and save predicted values –> y-hat.
  4. Estimate Model 2 (linear x) including y-hat (predicted y from Model 1) as an additional explanatory variable in the original model.
  5. Hypothesis pair: Null: parameter on y-hat = 0.
    Alternative: parameter on y-hat is not equal to zero.
  6. If null hypothesis is accepted, then the linear model is favored, i.e. the logarithmic specification didn’t add any explanatory power to model.
  7. Run the test again, this time assuming Model 1 (ln x) is correct.
51
Q
  1. When is the Ramsey RESET test used?
A

Mis-specification of functional form (MIS-specification: testing the null against an alternative that does not indicate what the correct specification is). Ramsey RESET is a test of linear specification against a non-linear specification. If include a quadratic term (to capture diminishing returns) in the restricted model, can test for higher powers using RESET test.

Linear model is the restricted model and non-linear model is the unrestricted model.

  1. Run regression and save fitted values of dependent variable.
  2. Add y-hat (fitted values from step 1 above) to the model (right-hand-side) and test its coefficient. If the T-stat of fitted variable (y-hat) from the unrestricted model is significant, then the model may be nonlinear.
  3. Alternatively the F-Stat or χ^2 can be used to test the restricted vs. unrestricted models:
    Null: Specification is linear.
    Alternative: Specification is NOT linear.
52
Q
  1. What is the dummy variable trap and how is it avoided?
A

Dummy variable trap is when create perfect multi-collinearity –> sum of all dummy variables is equal to 1 which is perfectly correlated w/ constant term.

Avoid the DV trap by omitting one of the dummy variable categories which becomes the reference group. Contrast is always w.r.t. the reference group. Each group needs enough observations to create precise contrasts.

53
Q
  1. How is the parameter on the dummy variable in OLS interpreted?
    e. g. dummy variable: 0 / 1, β-hat = -0.35.
A

Use equation: 100%[exp(β-hat) - 1] = 100%[exp(-0.35) - 1]
= 0.70 - 1 = -0.30.
Interpret as: The effect of switching from zero to unity is -0.30.

54
Q
  1. Describe threshold effects using the following model

income = β(1) + β(2)age + γ(B)Bachelor + γ(M)Master + γ(P)Phd + ε

A

β1 and β2 is the reference group -> high school diploma only.

β(1) + beta(2)age + γ(P)Phd –> Contrast: PhD over HS

β(1) + β(2)age + γ(B)Bachelor + γ(M)Master + γ(P)Phd –> Contrast: Ph.D over HS + B + M).

55
Q
  1. What test can be used to test if model w/dummy variables can be pooled?
A

Wald test for pooling (homework #7). Test stat is F-Stat or χ^2.
Null: Parameter estimates on terms including dummy variables = 0.
Null: Data can be pooled.
Alternative: One or more parameter estimates …… are not equal to 0.
Alternative: Data cannot be pooled.

If null is accepted, data can be pooled.
(p-value < α –> reject null)

56
Q
  1. What test can be used to confirm structural breaks?
A

ex post Chow Forecast Test (Homework #7)

Restricted model: Pools periods. Run the Chow Forecast Test indicating breakpoint between periods.

Null: Model is stable across periods.
Alternative: Model is not stable across periods.

If accept the null, can pool the periods.

57
Q
  1. Name four different types of non-linearities (in the variables).
A
  1. piecewise linear regression (use of structural breaks)
  2. log linear models
  3. models w/polynomials
  4. models w/ interaction effects
58
Q
  1. What test can be used for suspected endogeneity?
A

Hausman Test.
Null: Regressor(s) suspected of endogeneity are NOT endogenous.
Alternative: Regressor(s) suspected of endogeneity ARE endogenous.

If fail to reject null (p > α), OLS is the preferred estimation method –> higher precision.
If reject null (endogeneity exists), can use (1) Instrumental Variables (IV) if have exactly identified case (this is the strict form of IV estimator) -or- (2) two-stage least squares (2SLS) if have either exactly identified case or over-identified case.

Exactly identified case: # of outside IVs = # of endog. variables.
Over identified: # of outside IVs > # of endog. variables.

59
Q
  1. When choosing IVs when 1+ variables are endogenous. What test is used for instrument relevance?
A

Regress the variable suspected of endogenity on the inside (exogenous variables) and outside [IV(s) for endogenous variable(s)] instruments.
Null: Slope parameters are = zero (Instruments are WEAK/not relevant)
Alternative: Not all slope parameters = 0. (Instruments are STRONG/relevant).

Fail to accept the null IF: F-Stat > 10.
**Weak instruments cause problems for inference.

60
Q
  1. Describe two-stage least squares (2SLS).
A
  1. Stage 1: Regress endogenous variable on inside and outside instruments using OLS.
  2. Stage 2: Run original model (w/endogenous variable) as OLS BUT replace the endogenous variable with the predicted values of the dependent variable run Stage 1, i.e. replace x with x-hat derived in Stage 1.
  3. Test for instrument relevance.
  4. Test for instrument exogeneity.
  5. If IVs are relevant and exogenous, test the Stage 2 model for regressor endogeneity using the Hausman Test.
61
Q
  1. How is instrument exogeneity tested in 2SLS?
A
  1. Save the residuals from Stage 2 OLS fit in 2SLS.
  2. Regress the residuals on the inside and outside instruments.

Null: Slope parameters on outside instruments are 0. (Instruments are exogenous)
Alternative: Not all are 0. (Instruments are no exogenos).

Test Stat: J = mF ~ χ^2(m).
F-stat is from test of null, m = # of outside IVs minus # of endogenous regressors. (Can’t do test if m = 0).

62
Q
  1. What test is used for heteroskedasticity?
A
  1. White Heteroskedasticity Test.
    Null: No heteroskedasticity.
    Alternative: Heteroskedasticity of unknown, general form
    Test Stat: Obs*R-Squared (χ^2 distribution).
  2. Harvey Test: Special case of what White covers
63
Q
  1. What changes can be made if have heteroskedasticity?
A
  1. Robust estimation (White heteroskedasticity-consistent standard errors and covariances) –> new way.
  2. GLS or FGLS
64
Q
  1. What tests can be used to test for autocorrelation?
A
  1. Durbin-Watson Statistic (for OLS but not FGLS) –> good for only 1st order correlation.
    If ~2 (ρ = 0): Probably no autocorrelation.
    If closer to 0 (ρ = +1): Increasing amounts of positive first-order autocorrelation.
    If closer to 4 (ρ = -1): Increasing amounts of negative first-order autocorrelation.
  2. Lagrange Multiplier (LM) Test –> Specify number of lags –> first-order (1), second-order (2). Lags are the # of restrictions.
    Null: No autocorrelation
    Alternative: Autocorrelation
    Test Stat: χ^2.
    If reject null, estimate w/ FGLS assuming first/second/…. autocorrel.
    Run LM on above FGLS to test if additional autocorrelation (lags: 2).

Alternative estimation process for autocorrelation is Newey-West.

65
Q
  1. What are OLS-White, GLS, FGLS and when are they used?
A

OLS-White for Heteroskedasticity.
GLS: Generalized least squares.
FGLS: Feasible Generalized least squares.

For OLS, Ω is the Identity matrix (σ^2 is the same for all disturbances); however, when have heteroskedasticity or autocorrelation this isn’t true. OLS-White, GLS, and FGLS don’t assume that Ω is equal to identity matrix.

OLS-White: OLS modified in how variance-covariance matrix of parameter estimates is calculated. Results in robust estimators.
GLS: Assumes know value of Ω matrix.
FGLS: Ω contains unknown parameters that must be estimated. FGLS is large-sample approach.

If switch to GLS from OLS gain efficiency: GLS, when Ω is not identity matrix, is BLUE while OLS is no longer BLUE.

66
Q
  1. Name two systems equations. Why are they used? What is a typical problem w/system estimators?
A
  1. Seemingly unrelated regressions (SUR)
  2. 3SLS (used to estimate “simultaneous equations model”)

Used when endogeneity and/or autocorrelation (?)

Problem: Wrong specification will have spillover effects.