Mixture Model Setup

Visser & Speekenbrink: Sec 2.1, 2.2

Definition

Mixture model models yy as a series of random independent draws from different states (distributions). A mixture model with NN states is defined by

  • The state densities: fif_i, i=1,...,Ni = 1,..., N

  • The state probabilities: πi\pi_i, i=1,...,Ni = 1,...,N

The actual outcome is modeled as

f(Y=y)=i=1Nπifi(Y=y)f(Y=y) = \sum_{i=1}^N\pi_i f_i(Y=y)

Parameters

θi\theta_i: parameters of the component distributions fif_i

θobs\theta_{obs}: parameters of the observation densities, which contain all θi\theta_i for i=1,...,Ni=1,...,N

θpr\theta_{pr}: parameters for the mixing proportions.

zz: covariates. Currently ignored since we typically can model the mixture model with remaining parameters. But technically, we can model component distribution as fi(ytθi,zt)f_i(y_t|\theta_i, z_t) and mixing proportions as p(St=iZt,θpr)p(S_t = i|Z_t, \theta_{pr}). This allows distributions and proportions to vary across observations. Detail discussed later.

Likelihood

L(θy1:T)=t=1Ti=1Nπifi(ytθi)\begin{equation*} L(\theta|y_{1:T}) = \prod_{t=1}^T \sum_{i=1}^N \pi_i f_i(y_t|\theta_i) \end{equation*}

This likelihood can be optimized by using the EM algorithm or direct maximization. During the actual maximization process, we often choose to maximize log-likelihood.

l(θy1:T)=t=1Tlog(i=1Nπifi(ytθi))\begin{equation*} l(\theta|y_{1:T}) = \sum_{t=1}^T \log \left( \sum_{i=1}^N \pi_i f_i(y_t|\theta_i) \right) \end{equation*}

Posterior Probabilities

The posterior probability is the probability that the state at observation tt given observation is yty_t

p(St=iyt)=πif(ytSt=i)j=1Nπjf(ytSt=j)\begin{equation*} p(S_t = i | y_t) = \frac{\pi_i f(y_t|S_t=i)}{\sum_{j=1}^N \pi_j f(y_t|S_t=j)} \end{equation*}

Last updated