Lecture 7 (MLE and limited dependent variables) Flashcards

1
Q

What are the advantages and disadvantages of using an LPM to study the binary dependent variable?

A
  • Advantages
    • It is possible to directly interpret $\beta$ (one unit of $x$ increases the prob of $y$ by ..%)
    • The IV approach is easier with LPM than with non-LPM
  • Disadvantages
    • We might predict probabilities outside the unit interval (we try to predict something non-linear with a linear model) since the LMP is not bounded. This comes from the fact that the marginal effect is constant over the range of $x.$
    • Also, the variance depends on $x$ (from Palmes lecture)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Set up the latent index model and derive Pr(y=1)

use

y* = x’\beta+u

where y* is a latent variable

A

See notion and problem set

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is Maximum likelihood?

A

In OLS we choose the parameters that minimize SSR. In ML, given our
output (assumptions regarding probability distribution generating this) we choose
the parameter values maximizing the probability of the observed outcome.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What are the four steps in ML-estimation?

A

There are four main steps in maximum-likelihood (ML) estimation

  1. Derive the joint PDF of the sample (using the Bernoulli distribution)
  2. Obtain the likelihood function
  3. Take log of the likelihood function to obtain the log-likelihood function
  4. Take the partial derivative of the log-likelihood function w.r.t the parameter of interest (e.g., $\beta$) to obtain the ML-estimator from the FOC
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Do all 4 steps MLE steps for a binary latent variable Y.

Lets say we have derived G(x’B)

A

See notion

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

When is ML the most efficient estimator?

A
  • ML is the most efficient estimator if the distributional assumptions are
    correct
  • But efficiency comes at the price of robustness because we make some parametric assumptions
  • If some part of the model is mis-specified, e.g., the normality assumption is
    not true, the ML estimator is inconsistent
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

With probit, when will the density for y|x be incorrect?

A

With probit, the density for $y|x$ will be incorrect if

  • the error term $v$ is not normally distributed
  • or the latent error is not independent from x
  • or the latent variable model is not linear.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

How do we interpret the marginal effect of probit or logit?

A

See notion

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Explain and formulate an M-estimator

A

In statistics, M-estimators are a broad class of extremum estimators for which the objective function is a sample average. Both OLS, non-linear least squares and maximum likelihood estimation are special cases of M-estimators. The statistical procedure of evaluating an M-estimator on a data set is called M-estimation. GMM is a subclass of M-estimators.

See notion for formulation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What can we say regarding consistency and the M-estimator?

A

Under the assumptions:

  1. Identification assumption: $\beta_0$ is the unique solution
  2. Law of large numbers
    • q(x,\beta) is a continuous function
    • The space is compact (closed and bounded)

If 1 & 2 holds,\hat\beta is consistent.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

How can we learn about the asymptotics of an M-estimator?

A

We can also show that M-estimators are asymptotically normally distributed.

We do this using the score function.

We basically have the same procedure and get the same results as with OLS. Remember that OLS is a type of M-estimand.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly