Week 12: Binary Outcome Models and Maximum Likelihood Estimation Flashcards

1
Q

How else can we evaluate estimators?

A

Evaluating their asymptotic properties.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What are asymptotic properties?

A

What happens when the sample size grows to infinity.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

List 2 facts that are true if the sample was to grow to infinity.

A
  • If n → ∞, our estimator should get closer and closer to the estimand (consistency)
  • If n → ∞, the sampling distribution of our estimator should get closer and closer to the normal distribution (asymptotic normality)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

List the 3 features of the OLS estimator.

A
  • Unbiased: in expectation it produces the true parameters (βs and σ^2), under certain conditions
  • Consistent: as n → ∞, we get increasingly close to the true parameters
  • Normally Distributed in large samples
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is the Gauss-Markov Theorem?

A

The OLS estimator is BLUE – the Best Linear Unbiased
Estimator.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

List 5 assumptions of the Gauss-Markov Theorem.

A
  • The true model is a linear function of the parameters
  • The observations are randomly sampled
  • No explanatory variable (X) is constant, and there is no ‘perfect multicollinearity’ (no X is a linear combination of other Xs)
  • The expected value of εi conditional on xi is 0 (zero conditional mean assumption). Rules out bias from omitted variables
  • The variance of the errors is constant (homoscedasticity)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly