1. Dynamic Topic Model
1. Static Topic Model
eg. LDA
$\beta_{1:K}$ K topics, multinomial distribution of fixed volcabulary
Generative Process:
- Choose topic proportions θ from a distribution over the (K − 1)-simplex, such as a Dirichlet.
- For each word:
- Choose a topic assignment $Z ∼ Mult(\theta)$
- Choose a word $W ∼ Mult(\beta_z)$.
assumption:
documents are drawn exchangeably from the same set of topics.
2. Dynamic Topic Model
assumption:
topics evolve from last time period topics
Generative Process:
-
Draw topics in a state space model that evolves with Gaussian noise $$\beta_t \beta_{t−1} ∼ N (\beta_{t−1}, \sigma^2I)$$ -
Draw document specific topic proportion $$\alpha_t \alpha_{t−1} ∼ N (\alpha_{t−1}, \delta^2I) $$ - For each document:
- Draw $\eta∼N(\alpha_t, a^2I)$
- For each word:
- Draw $Z ∼ Mult(\pi(\eta))$.
- Draw $W_{t,d,n} ∼ Mult(\pi(\beta_{t,z}))$.
- ($pi(x) = \frac{exp(x)}{\sum exp(x)}$)
2. Approximate Inference
1. Static Topic Model
Gibbs Sampling
2. Dynamic Topic Model
Con: nonconjugacy of the Gaussian and multinomial models
variational methods:
optimize the free parameters of a distribution over the latent variables so that the distribution is close in Kullback-Liebler (KL) divergence to the true posterior; this distribution can then be used as a substitute for the true posterior.
1. Kalman Filter
TBC
2. Wavelet Regression
TBC