Gaussian mixture models structure datapoints
Info
Gaussian mixture models are a form of soft clustering as opposed to the hard clustering in ๐ K-Means Clustering.
Each cluster, represented by a ๐ Gaussian distribution, is defined by a centroid
In a generative sense, our data is generated from
where
We always optimize centroids
Info
Note that this equation is incredibly similar to ๐ถ Naive Bayes. If Naive Bayes lets
be a Gaussian distribution (instead of discrete), we get a Gaussian Mixture with independent (diagonal covariance ).
Training
Gaussian mixtures uses the ๐ Expectation Maximization algorithm to optimize its mixtures. We first find the mixture distribution
Algorithm
Given training data
Alternate until convergence.
- For each data point
, estimate . - Calculate new parameters for each mixture
Info
Intuitively, the E-step calculates cluster assignments, and the M-step finds the most likely parameters based off the assignments.
Prediction
Given point