EM-Algorithm Flashcards
What is the general idea if the EM-algorithm
- fill in missing values by estimated values
- estimate parameters given the filled in values
- re-estimate missing values with new parameter estimates
- re-estimate parameters etc….
iterate until convergence of parameters
What are the missing data mechanisms taken into account by EM-algorithm
MAR and MCAR
What does the EM-algorithm consist of?
Expectation step and a maximisation step
One advantage and one disadvantage of EM-algorithm
+ will reliably converge
- rate of convergence can be too slow when large fraction of missing information
What does the E-step do?
Calculates the conditional expectation of complete data log likelihood given the observed data and current estimates of parameters
What does the M-step do?
Performs maximum likelihood estimation on the (conditional) expected log likelihood found in E-step
Formal definition of the E-step
Q(θ, θ(t)) = Eymis[ l(θ | Yobs, Ymis) | Yobs, θ(t) ]
Formal definition of the M-step
θ(t+1) = arg max(θ) Q(θ | θ(t)
What does the GEM-algorithm do differently?
θ(t+1) not chosen to globally maximise Q(θ | θ(t) ) but rather to ensure Q(θ(t+1) | θ(t) ) is greater than Q(θ(t-1) | θ(t))
Explain stability / monotonicity in terms of EM-algorithm
Each step in the EM-algorithm leads to an observed likelihood that is greater than or equal to the previous likelihood
log ( L( θ(t+1) | Yobs) ) ≥ log ( L( θ(t) | Yobs) )