* Intro
wedge Admin
* Mid-Term Evaluation open
* remember to come by office hours once
wedge Recap
* From Rabiner, "A Tutorial on Hidden Markov Models and Selected Applications in Speech Recognition"
* Defining an HMM
wedge HMM is characterized by
wedge N, number of states in the model
* S, {S_1,S_2,S_3}
* state at time t is q_t
wedge M the number of distinct observation symbols per state
* V = {v_1,V_2,V_3}
wedge State transition probability
* A={a_ij}
* a_ij = P(q_t+1 = S_j | q_t = S_i) 1<= i, j <= N
* if all states can reach all states then a_ij > 0 for all i,m
wedge The Observation symbol probability distribution in state j
* B = {b_j(k)}
* b_j(k) = P(v_k at t | q_t = S_j) 1<= j <=N , 1<=k <=M
wedge The initial state distribution pi = {pi_i}
* pi_i = P(q_1=S_i) 1<= i <=N
* Example of how to use it generatively
* Model can be completely determined by M, N, A, B, and pi
wedge New Material
* Motivate with a house and motion sensor example
wedge 3 problems from Rabiner
wedge The 3 basic HMM problems
*
Pasted Graphic
Pasted Graphic

wedge Problem 1 is the evaluation problem
* It is also considered the scoring problem
* Choosing among multiple models
wedge Problem 2 is an attempt to uncover the hidden variable or find the correct state sequence
* "correct" is not accurate -> some optimality criterion there are several possible
wedge Problem 3 is the learning problem
* "training" the HMM based on some observed data
wedge Solving Problem #1
*
Pasted Graphic
Pasted Graphic

*
Pasted Graphic
Pasted Graphic

*
Pasted Graphic
Pasted Graphic

*
Pasted Graphic
Pasted Graphic

*
Pasted Graphic
Pasted Graphic

wedge A more tractable version is called the Forward-Backward procedure
*
Pasted Graphic
Pasted Graphic

*
Pasted Graphic
Pasted Graphic

*
Pasted Graphic
Pasted Graphic

*
Pasted Graphic
Pasted Graphic

*
Pasted Graphic
Pasted Graphic

*
Pasted Graphic
Pasted Graphic

wedge So this only describes the "forward variable"
* It is sufficient for problem 1
* but we will want to use a backward variable for the other problems
*
Pasted Graphic
Pasted Graphic

*
Pasted Graphic
Pasted Graphic

*
Pasted Graphic
Pasted Graphic

*
Pasted Graphic
Pasted Graphic

*
Pasted Graphic
Pasted Graphic

wedge Problem 2
wedge Lots of possible solutions
* "optimal"
wedge choose the states which are individually most likely
*
Pasted Graphic
Pasted Graphic

*
Pasted Graphic
Pasted Graphic

*
Pasted Graphic
Pasted Graphic

*
Pasted Graphic
Pasted Graphic

wedge Problems with this formulation
* State sequence may not be possible
* Other ways of optimizing are possible
wedge One well-known optimization is to optimize the single best state sequence
* Based on dynamic programming
* "Viterbi Algorithm"
*
Pasted Graphic
Pasted Graphic

*
Pasted Graphic
Pasted Graphic

*
Pasted Graphic
Pasted Graphic

*
Pasted Graphic
Pasted Graphic

*
Pasted Graphic
Pasted Graphic

*
Pasted Graphic
Pasted Graphic

wedge Two points
* This is very similar to the solution to Problem 1 except that we are maximizing rather than summing
* and it should be clear that it can be done efficiently with a lattice structure
wedge Problem 3
* Adusting the model parameters (A,B,pi) to maximize the probability of the observation sequence
wedge This is a global optimization problem
* No known analytical problem of maximize
wedge Use a local optimazation procedure called Baum-Welch algorithm
* iterative
* Expectation-Maximization procedure
*
*
Pasted Graphic
Pasted Graphic

*
Pasted Graphic
Pasted Graphic

*
Pasted Graphic
Pasted Graphic

*
Pasted Graphic
Pasted Graphic

*
Pasted Graphic
Pasted Graphic

*
Pasted Graphic
Pasted Graphic

*
Pasted Graphic
Pasted Graphic