Path: blob/master/notebooks/misc/dp_mixgauss_pos_truncat.ipynb
1192 views
Applying Dirichlet-processes to mixture-model clustering
(transformed from https://ericmjl.github.io/dl-workshop/04-gaussian-clustering/03-dirichlet-process-clustering.html )
We start with a data set from a Gaussian mixture distribution with unclear number of modes.
When samples are from the 1-d space, we can roughly read the number of components in the mixture distribution from the histogram of the samples:
However, even in the 1-d space, it is not always obvious to read the number of modes in a mixture distribution from the histogram. The information starts blurring as we increase the number of bins in the histogram.
Where the number of clusters is not clear, we can turn to DP mixture models, which does not impose an exact number of clusters, but controls the number of clusters probabilistically with a single concentration parameter.
Here we fit a DP Gaussian mixture model (DP-GMM) to above data and learn
the concentration parameter of the DP,
the optimal relative weighting of components, conditioned on concentration parameters, and
distribution paraemters for each component, conditioned on the data.
We define the Bayesian model and try to obtain the maxinum a posterior (MAP) estimation of the these parameters. This is equivalent to miaximize the joint likelihood of these parameters and the data. Since we do not assign a prior distribution for the concentration parameter of the DP, the estimation of the concentration parameter can be seen as an empirical Bayesian approach.