Presentation is loading. Please wait.

Presentation is loading. Please wait.

Forward-backward algorithm LING 572 Fei Xia 02/23/06.

Similar presentations


Presentation on theme: "Forward-backward algorithm LING 572 Fei Xia 02/23/06."— Presentation transcript:

1 Forward-backward algorithm LING 572 Fei Xia 02/23/06

2 Outline Forward and backward probability Expected counts and update formulae Relation with EM

3 HMM A HMM is a tuple : –A set of states S={s 1, s 2, …, s N }. –A set of output symbols Σ={w 1, …, w M }. –Initial state probabilities –State transition prob: A={a ij }. –Symbol emission prob: B={b ijk } State sequence: X 1 …X T+1 Output sequence: o 1 …o T

4 Constraints

5 Decoding Given the observation O 1,T =o 1 …o T, find the state sequence X 1,T+1 =X 1 … X T+1 that maximizes P(X 1,T+1 | O 1,T ).  Viterbi algorithm X1X1 X2X2 XTXT … o1o1 o2o2 oToT X T+1

6 Notation A sentence: O 1,T =o 1 …o T, T is the sentence length The state sequence X 1,T+1 =X 1 … X T+1 t: time t, range from 1 to T+1. X t : the state at time t. i, j: state s i, s j. k: word w k in the vocabulary

7 Forward and backward probabilities

8 Forward probability The probability of producing o i,t-1 while ending up in state s i :

9 Calculating forward probability Initialization: Induction:

10 Backward probability The probability of producing the sequence O t,T, given that at time t, we are at state s i.

11 Calculating backward probability Initialization: Induction:

12 Calculating the prob of the observation

13 Estimating parameters The prob of traversing a certain arc at time t given O: (denoted by p t (i, j) in M&S)

14 The prob of being at state i at time t given O:

15 Expected counts Sum over the time index: Expected # of transitions from state i to j in O: Expected # of transitions from state i in O:

16 Update parameters

17 Final formulae

18 Emission probabilities Arc-emission HMM:

19 The inner loop for forward-backward algorithm Given an input sequence and 1.Calculate forward probability: Base case Recursive case: 2.Calculate backward probability: Base case: Recursive case: 3.Calculate expected counts: 4.Update the parameters:

20 Relation to EM

21 HMM is a PM (Product of Multi-nominal) Model Forward-back algorithm is a special case of the EM algorithm for PM Models. X (observed data): each data point is an O 1T. Y (hidden data): state sequence X 1T. Θ (parameters): a ij, b ijk, π i.

22 Relation to EM (cont)

23

24 Iterations Each iteration provides values for all the parameters The new model always improve the likeliness of the training data: The algorithm does not guarantee to reach global maximum.

25 Summary A way of estimating parameters for HMM –Define forward and backward probability, which can calculated efficiently (DP) –Given an initial parameter setting, we re-estimate the parameters at each iteration. –The forward-backward algorithm is a special case of EM algorithm for PM model

26 Additional slides

27 Definitions so far The prob of producing O 1,t-1, and ending at state s i at time t: The prob of producing the sequence O t,T, given that at time t, we are at state s i : The prob of being at state i at time t given O:

28

29 Emission probabilities Arc-emission HMM: State-emission HMM:


Download ppt "Forward-backward algorithm LING 572 Fei Xia 02/23/06."

Similar presentations


Ads by Google