In order to fit finite Bayesian mixture models, the FMM procedure treats the mixture model as a missing data problem and introduces an assignment variable as in Dempster, Laird, and Rubin (1977). Since is not observable, it is frequently referred to as a latent variable. The unobservable variable assigns an observation to a component in the mixture model. The number of states, k, might be unknown, but it is known to be finite. Conditioning on the latent variable , the component memberships of each observation is assumed to be known, and Bayesian estimation is straightforward for each component in the finite mixture model. That is, conditional on , the distribution of the response is now assumed to be . In other words, each distinct state of the random variable leads to a distinct set of parameters. The parameters in each component individually are then updated using a conjugate Gibbs sampler (where available) or a Metropolis-Hastings sampling algorithm.
The FMM procedure assumes that the random variable has a discrete multinomial distribution with probability of belonging to a component j; it can occupy one of k states. The distribution for the latent variable is
where denotes a conditional probability density. The parameters in the density denote the probability that S takes on state j.
The FMM procedure assumes a conjugate Dirichlet prior distribution on the mixture proportions written as:
|
|
|
where indicates a prior distribution.
Using Bayes’ theorem, the likelihood function and prior distributions determine a conditionally conjugate posterior distribution of and from the multinomial distribution and Dirichlet distribution, respectively.