Gaussian Mixture Models
- A probabilistic model
- Assumes all data points are generated from a mixture of finite no. of gaussian
distributions - The parameters of the gaussian distributions are unknown.
- It is a way of generalizing k-mean(or k-medoid or k-mode for that matter) clustering to use the
co-variance structure/stats as well as the mean/central-tendency measures of latent
gaussians.
scikit-learn
- Implements the Expectation Maximization algorithm.
Pros:
- Fastest for learning mixture models
- No bias of means towards zero, or bias cluster sizes to have specific structures
Cons:
- When there’s not enough points per mixture, estimating covariance matrices becomes
difficult -
Number of components; will always use all the components it has access to, so might need
missing or test-reserved data.. -
No. of components can be chosen based on BIC criterion.
- Variational Bayesian Gaussian mixture avoids having to specify number of components
Variational Bayesian Gaussian Mixture
- — Uses Variational Inference algorithm to estimate