Lecture 11 - Clustering: EM Algorithm for GMM Flashcards
What is a Gaussian Mixture Model (GMM)?
A probabilistic model representing data as a mixture of Gaussian distributions.
Priors - 𝜋𝑖
𝜋𝑖 : Probabilities of each cluster.
Means - 𝜇𝑖
𝜇𝑖: Centers of Gaussian clusters.
Covariances - Σ𝑖
Σ𝑖: Shapes and spreads of clusters.
Why use GMM instead of K-Means?
GMM vs. K-Means
Soft Clustering: Assigns probabilities to clusters, not hard assignments.
Handles Arbitrary Shapes: Covariances allow ellipsoid-shaped clusters.
Probabilistic Reasoning: More robust for complex data distributions.
Mnemonic: “GMM Gives More Meaningful shapes.”
What is the Expectation-Maximization (EM) Algorithm?
An iterative algorithm to optimize parameters (π,μ,Σ) in GMM
What happens in the E-Step?
Compute the posterior probability hi,t that a point belongs to a cluster:
What happens in the M-Step?
Update parameters using hi,j.
How does EM converge?
Converges when the log-likelihood stabilizes.
Alternative: Stop when the change in log-likelihood is below a threshold.
Tip: Imagine EM climbing a mountain of likelihood to the peak (maximum).
Challenges with GMM and EM
Initialization Sensitivity: Poor starts may lead to local optima.
Choosing 𝑘: Methods include Elbow, BIC, and AIC.
Computational Cost: Iterative nature makes it slow for large datasets.
Applications of GMM. Where is GMM used?
Image Segmentation: Dividing an image into meaningful regions.
Speech Recognition: Modeling phonemes or accents.
Anomaly Detection: Identifying outliers with low probabilities.