Mixture models¶
The simplest form of LVM is when \(z_i \in \{1, \cdots,K\}\), representing a discrete latent state. We use a discrete prior \(p(z_i) = Cat(\pi)\). The likelihood function we use \(p(x_i,z_i = k) = p_k(x_i)\) where \(p_k\) is the k’th base distribtuion for the observations (it can be of any type). The overall model is known as a mixture model, since we mix togeter the K base distributions as follows:
Which is a convex combination of the \(p_k\)’s where we are taking a weighted sum, where the mixin weights \(\pi_k\) satisfy \(0 \le \pi_k \le 1\) and \(\sum_{k=1}^K \pi_k =1\).
The following tables gives an example of different mixture models:
Name |
Likelihood |
Prior |
---|---|---|
Mixture of Gaussians |
MVN |
Discrete |
Mixture of multinomials |
Prod. Discrete |
Discrete |
Factor analysis/ Probabilistic PCA |
Prod.. Gaussian |
Prod. Gaussian |
Probabilistic ICA |
Prod. Gaussian |
Prod. Laplace |
Multinomial PCA |
Prod. Discrete |
Prod. Gaussian |
Latent Dirichlet allocation |
Prod. Discrete |
Dirichlet |