Fall 2025, Prof Sarah Dean
"What we do"
"Why we do it"
Consider a hidden Markov model with transitions \(P\) and observation probabilities \(B\) $$ S_{t+1} \sim P^\top e_{s_t},\quad Y_t \sim B^\top e_{s_t} $$
Fact: the filtering algorithm computes the posterior $$ \mathbb P\{S_{t}=i|y_0,...,y_t\}=\mu_{t|t}[i] $$
Fact: the relationship between the observations and the estimated state is bilinear, so predictions depend polynomially on past observations
Reference: Chapter 17 of Machine Learning: A Probabilistic Perspective
Next time: generalization and online learning