Final Exam Flashcards
The classical assumptions must be
met in order for OLS estimators to be the best available
Classical Assumption #1
The regression model is linear, is correctly specified, and has an additive error term
Classical Assumption #2
The error term has a zero population mean
Classical Assumption #3
All explanatory variables are uncorrelated with the error term
Classical Assumption #4
Observations of the error term are uncorrelated with each other (no serial correlation)
Classical Assumption #5
The error term has a constant variance (no heteroskedasticity)
Classical Assumption #6
No explanatory variable is a perfect linear function of any other explanatory variable(s) (no perfect multicollinearity)
Classical Assumption #7
The error term is normally distributed
Omitted Variable Bias (Conditions)
relevant (β2 ≠ 0) – X1 and X2 are correlated
Expected bias
Expected bias in መ 𝛽1 has two components: the sign of β2 and the sign of Corr(X1, X2)
Limited Dependent Variables
We have discussed dummy variables (indicator variables, binary variables) as a tool for measuring qualitative/categorical independent variables (gender, race, etc.)
linear probability model
simply running OLS for a regression, where the dependent variable is a dummy (i.e. binary) variable:
where Di is a dummy variable, and the Xs, βs, and ε are typical independent variables, regression coefficients, and an error term, respectively
the term, linear probability model
comes from the fact that the right side of the equation is linear while the expected value of the left side measures the probability that Di = 1
Some issues with LPM
𝑌 𝑖 ≤ 0 or 𝑌 𝑖 ≥ 1 a more fundamental problem with the linear probability model: nothing in the model requires 𝑌 to be between 0 and 1! If 𝑌 is not between 0 and 1, how do we interpret it as a probability? A related limitation is that the marginal effect of a 1-unit increase in any X is forced to be constant, which cannot possibly be true for all values of X E.g., if increasing X by 1 always increases Y by a particular amount, 𝑌 must exceed 1 when X is sufficiently large
The Binomial Logit Model
The binomial logit is an estimation technique for equations with dummy dependent variables that avoids the unboundedness problem of the linear probability model
Logits cannot be estimated using OLS
but are instead estimated by maximum likelihood (ML), an iterative estimation technique that is especially useful for equations that are nonlinear in the coefficients Again, for the logit model is bounded by 1 and 0
Basic Procedure for Random Assignment Experiments
Recruit sample of subjects Randomly assign some to treatment group and some to control group ◼ Random assignment makes treatment uncorrelated with individual characteristics Measure average difference in outcomes between treatment and control groups
natural experiments (or quasi-experiments
attempt to utilize the “treatment-control” framework in the absence of actual random assignment to treatment and control groups
Difference-in-difference estimator:
Policy impact = (Tpost – Tpre) – (Cpost– Cpre) T: treatment group outcome, C: control group outcome The DD estimate is the amount by which the change for the treatment group exceeded the change for the control grou
Panel data:
repeated observations of multiple units over time (combination of cross-sectional and time-series)
Main advantages of panel data
Increased sample size Ability to answer types of questions that cross-sectional and time-series data cannot accommodate Enables use of additional methods to eliminate omitted variables bias
Panel Data Notation
𝑖 subscript indexes the cross-sectional unit (individual, county, state, etc.) t subscript indexes the time period in which the unit is observe
First-differenced estimator
∆𝑌 𝑖 = 𝛼0 + 𝛽1∆𝑋𝑖 + ∆𝜀𝑖
Advantages of random effects estimator (if assumption about 𝑎𝑖 is correct):
Allows time-invariant regressors to be included More degrees of freedom (only estimates parameters of distribution from which 𝑎𝑖 is assumed to be drawn; fixed effects estimator uses one degree of freedom per fixed effect)
Disadvantages of random effects estimator:
Biased if assumption that 𝑎𝑖 is uncorrelated with regressors is incorrect (while FE estimator allows arbitrary correlation between 𝑎𝑖 and regressors)
❑ Fixed effects estimator widely preferred when regressors of interest are time-varying
It rarely seems likely that 𝑎𝑖 is uncorrelated with any regressors; fixed effects model is generally far more convincing
Hausman test
❑ Fixed and random effect estimators can be compared with a Hausman test (previously seen in instrumental variables context as test for endogeneity)
Fixed vs. random effects - General Advice
Under random effects hypothesis, both RE and FE estimators are consistent (should give similar results); under alternative hypothesis, FE consistent but RE is not Therefore, if estimates are significantly different, can reject null hypothesis of random effects
Fixed vs. random effects - Concept
❑ General advice: use fixed effects estimator if it’s feasible
T-test
Divide the coefficient by the standard error to get the t-value
Omitted Variables – Bias Assessment
Sign (β2) * Sign (Corr (X1, X2)) = Sign of Bias (β1)
Irrelevant Variables - Inclusion Criteria
Theory: is there sound justification for including the variable?
Bias: do the coefficients for other variables change noticeably when the variable is included?
T-Test: is the variable’s estimated coefficient statistically significant?
R-square: has the R-square (adjusted R-square) improved?