In statistics, a mixture model is a probabilistic model for representing the presence of subpopulations within an overall population, without requiring that an observed data set should identify the sub-population to which an individual observation belongs. - Wikipedia
Gaussian Mixture Models (GMMs)
GMM: Superposition of K Gaussian densities of the following form is mixture of Gaussians.
Mixture coefficients:
We get the bottom relation by integrating both sides on the above equation.
Likelihood Function:
where
Log Likelihood:
This is a hard problem.
E-step
- compute soft assignment (posterior probabilities)
M-step
- re-estimate parameters
Advantages
- Flexibility, K-Means just assumes that clusters are spherical.
- Uncertainty estimation is an added benefit of soft assignment.
- Density Estimate, can help in identify outliers, anomalies (low
) - Useful as a Generative Model
- Less sensitive to initialization than K-Means
How to choose K?
- Pick the 'k' which generates maximum likelihood for a 'hold out' set.
- Cross-validation, Information Criteria (AIC, BIC)