0

I have read a couple of references on the interpretation of a latent variable in the GMM/EM literature and I found a massive discrepancy between the authors so much so I now have no idea how GMM/EM works.

Here is the first set of definition:

The latent variable is a discrete random variable that takes on values $z \in \{1,\ldots, K\}$ and is said to have categorical distribution, where $P[z = k] = \pi_k$

See page 26: https://www.cs.toronto.edu/~urtasun/courses/CSC411_Fall16/13_mog.pdf

https://ermongroup.github.io/cs228-notes/learning/latent/

http://www2.stat.duke.edu/~sayan/Sta613/2018/lec/emnotes.pdf

https://arxiv.org/pdf/1712.01263.pdf

And this question: Latent variable in Gaussian Mixture Model

Here is the second set of definition:

The latent variable is a discrete random variable $z$ that is a one-hot encoded vector, e.g., [1,0,0], [0,1,0], [0,0,1], where each component of $z = (z_k)$ is an indicator variable $z_k \in \{0, 1\}$ and we write $P[z_k = 1] = \pi_k$

See: M. P. Deisenroth, A. A. Faisal, C. S. Ong, "Mathematics of Machine Learning" page 363-364

Bishop: page 39 of these notes https://www2.cs.sfu.ca/~oschulte/teaching/726/spring11/slides/mychapter9.pdf

http://pillowlab.princeton.edu/teaching/statneuro2018/slides/notes16_LatentsAndEM.pdf

https://arxiv.org/pdf/2001.02923.pdf

And this question Question about the latent variable in EM algorithm


So the problem to me is that sometimes people are writing $P[z = k] = \pi_k$, and other times people are writing $P[z_k = 1] = \pi_k$.

What is going on?

0 Answers0