3) Maximum Likelihood Estimation Flashcards
What is the Likelihood Function and Log-Likelihood Function
What is the score function
The first derivative of the log-likelihood function with regard to the parameters
What steps are involved in finding the Maximum Likelihood Estimate (MLE)
- The condition for MLE is πΊπ(π½Λ ππΏ) = 0 indicating the point where the log-likelihood functionβs slope is zero, suggesting a maximum or minimum.
- Confirming a Maximum:
- Scalar ΞΈ: Ensure the second derivative of the log-likelihood is negative confirming maximum at π½ΛππΏ
- Vector ΞΈ: Compute the Hessian matrix at the MLE. The MLE is confirmed if this matrix is negative definite (all eigenvalues are negative), ensuring the log-likelihood is locally concave.
How can cross-entropy and KL divergence be approximated for large sample sizes
For large sample size n, the cross-entropy π»(πΉ, πΊπ½) is approximated by the empirical cross-entropy where the expectation is taken with regard to πΉΛπ rather than πΉ
How is maximizing the log-likelihood related to minimizing KL divergence for large sample sizes
What does it mean for an estimator to be consistent
Does the maximum likelihood function produce consistent estimates
Yes the method of maximum likelihood in general produces consistent estimates
Why are MLEs considered asymptotically unbiased
Maximum Likelihood Estimates (MLEs) are considered asymptotically unbiased because, with sufficient data, the MLE will converge to the true parameter value
How does maximum likelihood estimation connect to least squares estimation in a normal distribution model
In a normal distribution model, finding the mean using maximum likelihood estimation is equivalent to using least squares estimation.
What is the observed Fisher information
The negative curvature at the MLE:
π±π(π½ΛππΏ), the derivative of the negative score function evaluated at the MLE
What is the relationship between observed and expected Fisher information