Mixture models¶
The simplest form of LVM is when \(z_i \in \{1, \cdots,K\}\), representing a discrete latent state. We use a discrete prior \(p(z_i) = Cat(\pi)\). The likelihood function we use \(p(x_i,z_i = k) = p_k(x_i)\) where \(p_k\) is the k’th base distribtuion for the observations (it can be of any type). The overall model is known as a mixture model, since we mix togeter the K base distributions as follows:
Which is a convex combination of the \(p_k\)’s where we are taking a weighted sum, where the mixin weights \(\pi_k\) satisfy \(0 \le \pi_k \le 1\) and \(\sum_{k=1}^K \pi_k =1\).
The following tables gives an example of different mixture models:
| Name | Likelihood | Prior | 
|---|---|---|
| Mixture of Gaussians | MVN | Discrete | 
| Mixture of multinomials | Prod. Discrete | Discrete | 
| Factor analysis/ Probabilistic PCA | Prod.. Gaussian | Prod. Gaussian | 
| Probabilistic ICA | Prod. Gaussian | Prod. Laplace | 
| Multinomial PCA | Prod. Discrete | Prod. Gaussian | 
| Latent Dirichlet allocation | Prod. Discrete | Dirichlet | 
