Maximum likelihood estimation Flashcards

1
Q

Log-likelihood for GLMs

A

θ = (β, σ2)
l(θ) = -n/2 ln2π - n/2lnσ2 -1/(2σ2) (y-Xβ)’(y-Xβ)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Maximum likelihood estimator for β

A

β^ = argmax(β ∈ Rp) l(β, σ2) = (X’X)-1X’y
Equivalent to OLS estimator

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Properties of β^

A
  • β^|X ~ Np(β, σ^2(X’X)-1));
  • Unbiased;
  • Efficient in the Cramer-Rao sense.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Maximum likelihood estimator for σ2

A

σ2^ML = argmax(σ2 ∈ R) l(β, σ2) = ε^’ε^ / n = SSE / n
Raw residuals ε^ = y - y^ = y - X(X’X)-1X’y = [In - X(X’X)-1X’] y = (In - H)y

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Unbiased estimator for σ2

A

σ2^= SSE / (n-p)
Also obtainable thorugh the restricted liklelihood function maximisation. It is unbiased and independent of β^ too.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Properties related to σ2^ML

A
  • ε^|X ~Nn(0, σ2(In - H) that can be standardized into ri = ε^i / √ σ2(1-Hii) and then r|X -d-> Nn(0, In)
  • SSE/σ2 ~ χ2n-p, which leads to E[σ2ML] = σ2 (n-p)/n so biased even in asimptotically unbiased
  • SSE/σ2 ⊥ β^ => σ2ML ⊥ β^
How well did you know this?
1
Not at all
2
3
4
5
Perfectly