Presentation is loading. Please wait.

Presentation is loading. Please wait.

Part 6 HMM in Practice CSE717, SPRING 2008 CUBS, Univ at Buffalo.

Similar presentations

Presentation on theme: "Part 6 HMM in Practice CSE717, SPRING 2008 CUBS, Univ at Buffalo."— Presentation transcript:

1 Part 6 HMM in Practice CSE717, SPRING 2008 CUBS, Univ at Buffalo

2 Practical Problems in the HMM Computation with Probabilities Configuration of HMM Robust Parameter Estimation (Feature Optimization, Tying) Efficient Model Evaluation (Beam Search, Pruning)

3 Computation with Probabilities Logarithmic Probability Representation Lower Bounds for Probabilities Codebook for Semi-Continuous HMMs

4 Probability of State Sequence s for a Given Model λ If all, for a sequence of T>100,

5 Logarithm Transformation

6 Kingsbury-Rayner Formula

7 Mixture Density Model Kingsbury-Rayner Formula is not advisable here (too many exps and logs) Approximation

8 Lower Bounds for Probabilities Choose a minimal probability For example: In training it is avoided that certain states are not considered for parameter estimation In decoding it is avoided that paths through states with vanishing output probabilities are immediately discarded

9 Codebook Evaluation for Semi- Continuous HMMs Semi-Continuous HMM

10 Codebook Evaluation for Semi- Continuous HMMs By Bayes’ Law Assume can be approximated by a uniform distribution, then

11 Codebook Evaluation for Semi- Continuous HMMs This reduces the dynamic range of all quantities involved

12 Configuration of HMM Model Topology Modularization Compound Models Modeling Emissions

13 Model Topology Input data of speech and handwriting recognition exhibit a chronological or linear structure Ergodic model is not necessary

14 Linear Model The most simple model that describes chronological sequences Transitions to the next state and to the current state are allowed

15 Bakis Model Skipping of states is allowed Larger flexibility inn the modeling of duration Widely used in speech and handwriting recognition

16 Left-to-right Model An arbitrary number of states may be skipped in forward direction Jumping back to “past” states is not allowed Can describe larger variations in the temporal structure; longer parts of the data may be missing

17 Modularization English Word Recognition Thousands of words: more than thousands of word models; requires large amount of training data 26 letters: limited number of character models Modularization: divides complex model into smaller models of segmentation units Word -> subword -> character

18 Variation of Segmentation Units in Different Context Phonetic transcription of word “speech” : /spitS/ Cannot easily be distinguished from achieve (/@tSiv/), cheese (/tSiz/), or reality (/riEl@ti/) Triphone [Schwartz, 1984] Three immediately neighboring phone units taken as a segmentation units, e.g., p/i/t Eliminates the dependence of the variability of segmentation units on the context

19 Compound Models Parallel connection of all individual word models HMM structure for isolated word recognition Circles: Model States Squares: Non-emission States HMM structure for connected word recognition

20 Grammar Coded into HMM

21 Modeling Emissions Continuous feature vectors in the fields of speech and handwriting recognition are described by mixture models Size of the codebook and number of component densities per mixture density need to be decided No general way; a compromise between the precision of the model, its generalization capabilities, and the computation time Semi-Continuous Model Size of codebook: some hundred up to a few thousand densities Mixture Model: 8 to 64 component densities

22 References [1] Schwartz R, Chow Y, Roucos S, Krasner M, Makhoul J, Improved hidden Markov Modelling of phonemes for continuous speech recognition, in International Conference on Acoustics, Speech and Signal Processing, pp 35.6.1-35.6.4, 1984.

23 Robust Parameter Estimation Feature Optimization Tying

24 Feature Optimization Techniques

Download ppt "Part 6 HMM in Practice CSE717, SPRING 2008 CUBS, Univ at Buffalo."

Similar presentations

Ads by Google