Machine Learning Hidden Markov Model Darshana Pathak University of North Carolina at Chapel Hill Research Seminar – November 14, 2012.

Slides:



Advertisements
Similar presentations
An Analysis of Hamptonese Using Hidden Markov Models Ethan Le and Mark Stamp Department of Computer Science San Jose State University McNair Scholars Program.
Advertisements

Hidden Markov Models By Marc Sobel. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 Introduction Modeling.
Automatic Speech Recognition II  Hidden Markov Models  Neural Network.
Hidden Markov Models Reading: Russell and Norvig, Chapter 15, Sections
HIDDEN MARKOV MODELS Prof. Navneet Goyal Department of Computer Science BITS, Pilani Presentation based on: & on presentation on HMM by Jianfeng Tang Old.
Hidden Markov Models. A Hidden Markov Model consists of 1.A sequence of states {X t |t  T } = {X 1, X 2,..., X T }, and 2.A sequence of observations.
Introduction to Hidden Markov Models
數據分析 David Shiuan Department of Life Science Institute of Biotechnology Interdisciplinary Program of Bioinformatics National Dong Hwa University.
Hidden Markov Models Ellen Walker Bioinformatics Hiram College, 2008.
Statistical NLP: Lecture 11
Ch-9: Markov Models Prepared by Qaiser Abbas ( )
Hidden Markov Models Theory By Johan Walters (SR 2003)
Hidden Markov Model Most pages of the slides are from lecture notes from Prof. Serafim Batzoglou’s course in Stanford: CS 262: Computational Genomics (Winter.
Statistical NLP: Hidden Markov Models Updated 8/12/2005.
Hidden Markov Models Fundamentals and applications to bioinformatics.
Hidden Markov Models in NLP
Hidden Markov Model based 2D Shape Classification Ninad Thakoor 1 and Jean Gao 2 1 Electrical Engineering, University of Texas at Arlington, TX-76013,
Lecture 15 Hidden Markov Models Dr. Jianjun Hu mleg.cse.sc.edu/edu/csce833 CSCE833 Machine Learning University of South Carolina Department of Computer.
Hidden Markov Models 1 2 K … 1 2 K … 1 2 K … … … … 1 2 K … x1x1 x2x2 x3x3 xKxK 2 1 K 2.
1. Markov Process 2. States 3. Transition Matrix 4. Stochastic Matrix 5. Distribution Matrix 6. Distribution Matrix for n 7. Interpretation of the Entries.
SPEECH RECOGNITION Kunal Shalia and Dima Smirnov.
Albert Gatt Corpora and Statistical Methods Lecture 8.
Presenter: Yufan Liu November 17th,
Probabilistic reasoning over time So far, we’ve mostly dealt with episodic environments –One exception: games with multiple moves In particular, the Bayesian.
Data Mining Techniques Outline
Hamptonese1 Hamptonese and Hidden Markov Models or When is a Cipher not a Cipher? Ethan Le and Mark Stamp Department of Computer Science San Jose State.
Metamorphic Malware Research
1 Hidden Markov Model Instructor : Saeed Shiry  CHAPTER 13 ETHEM ALPAYDIN © The MIT Press, 2004.
Probabilistic Model of Sequences Bob Durrant School of Computer Science University of Birmingham (Slides: Dr Ata Kabán)
Bayesian Networks Alan Ritter.
CS 188: Artificial Intelligence Fall 2009 Lecture 19: Hidden Markov Models 11/3/2009 Dan Klein – UC Berkeley.
CS6800 Advanced Theory of Computation Fall 2012 Vinay B Gavirangaswamy
Introduction to Profile Hidden Markov Models
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
A Revealing Introduction to Hidden Markov Models
Isolated-Word Speech Recognition Using Hidden Markov Models
CSCE555 Bioinformatics Lecture 6 Hidden Markov Models Meeting: MW 4:00PM-5:15PM SWGN2A21 Instructor: Dr. Jianjun Hu Course page:
THE HIDDEN MARKOV MODEL (HMM)
BINF6201/8201 Hidden Markov Models for Sequence Analysis
A Revealing Introduction to Hidden Markov Models
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
Hidden Markov Models in Keystroke Dynamics Md Liakat Ali, John V. Monaco, and Charles C. Tappert Seidenberg School of CSIS, Pace University, White Plains,
1 Generative and Discriminative Models Jie Tang Department of Computer Science & Technology Tsinghua University 2012.
CS774. Markov Random Field : Theory and Application Lecture 19 Kyomin Jung KAIST Nov
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: ML and Simple Regression Bias of the ML Estimate Variance of the ML Estimate.
Processing Sequential Sensor Data The “John Krumm perspective” Thomas Plötz November 29 th, 2011.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
© 2015 McGraw-Hill Education. All rights reserved. Chapter 19 Markov Decision Processes.
Theory of Computations III CS-6800 |SPRING
1 CONTEXT DEPENDENT CLASSIFICATION  Remember: Bayes rule  Here: The class to which a feature vector belongs depends on:  Its own value  The values.
CSE 5331/7331 F'07© Prentice Hall1 CSE 5331/7331 Fall 2007 Machine Learning Margaret H. Dunham Department of Computer Science and Engineering Southern.
Context-based vision system for place and object recognition Antonio Torralba Kevin Murphy Bill Freeman Mark Rubin Presented by David Lee Some slides borrowed.
 Present by 陳群元.  Introduction  Previous work  Predicting motion patterns  Spatio-temporal transition distribution  Discerning pedestrians  Experimental.
Dongfang Xu School of Information
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Elements of a Discrete Model Evaluation.
Hidden Markov Models Sean Callen Joel Henningsen.
1 Hidden Markov Models Hsin-min Wang References: 1.L. R. Rabiner and B. H. Juang, (1993) Fundamentals of Speech Recognition, Chapter.
Stochastic Processes and Transition Probabilities D Nagesh Kumar, IISc Water Resources Planning and Management: M6L5 Stochastic Optimization.
Hidden Markov Models. A Hidden Markov Model consists of 1.A sequence of states {X t |t  T } = {X 1, X 2,..., X T }, and 2.A sequence of observations.
Definition of the Hidden Markov Model A Seminar Speech Recognition presentation A Seminar Speech Recognition presentation October 24 th 2002 Pieter Bas.
Other Models for Time Series. The Hidden Markov Model (HMM)
Hidden Markov Models Sean Callen Joel Henningsen.
By: Nicole Cappella. Why I chose Speech Recognition  Always interested me  Dr. Phil Show Manti Teo Girlfriend Hoax  Three separate voice analysts proved.
Partially Observable Markov Decision Process and RL
A Revealing Introduction to Hidden Markov Models
An INTRODUCTION TO HIDDEN MARKOV MODEL
CSCI 5822 Probabilistic Models of Human and Machine Learning
Hidden Markov Autoregressive Models
HCI/ComS 575X: Computational Perception
CONTEXT DEPENDENT CLASSIFICATION
Presentation transcript:

Machine Learning Hidden Markov Model Darshana Pathak University of North Carolina at Chapel Hill Research Seminar – November 14, 2012

Disclaimer All the information in the following slides assumes that “There is a GREAT human mind behind every computer program.”

What is Machine Learning? Make Computers learn from a given task and experience. “Field of study that gives computers the ability to learn without being explicitly programmed”. - Arthur Samuel (1959)Arthur Samuel

Why Machine Learning? Human Learning is terribly slow! (?) o 6 years to start school, around 20 more years to become cognitive /computer scientist... o Linear programming, calculus, Gaussian models, optimization techniques and so on…

Why Machine Learning? No copy process in human beings - ‘one-trial learning’ in computers. Computers can be programmed to learn – Both human and computer programs make errors, error is predictable for computer, we can measure error.

Some more reasons… Growing flood of electronic data – Machines can digest huge amounts of data which is not possible for human. Supporting computational power is also growing! Data mining – to help improve decisions o Medical records study for diagnosis o Speech/handwriting/face recognition o Autonomous driving, robots

Important Distinction Machine learning focuses on prediction, based on known properties learned from the training data. Data mining focuses on the discovery of (previously) unknown properties on the data. Example: Purchase history/behavior of a customer.

Hidden Markov Model

Hidden Markov Model - HMM A Markov model with hidden states. Markov Model – Stochastic Model that assumes Markov property. Stochastic model – A system with stochastic process (random process).

HMM – Stochastic model Stochastic process vs. Deterministic process. – SP is probabilistic counterpart of DP. Examples: – Games involving dice and cards, coin toss. – Speech, audio, video signals – Brownian motion – Medical data of patients – Typing behavior (Related to my project)

HMM – Markov Model Markov Model – Stochastic Model that assumes Markov property. Markov property  Memory-less property – Future states of the process depend only upon the present state, – And not on the sequence of events that preceded it.

Funny example of Markov chain 0 – Home; 4 – Destination 1,2,3 corners;

Hidden Markov Model - HMM A Markov model with hidden states – Partially observable system. Simple Markov ModelHidden Markov Model Every state is directly visible to the observer. The state is not directly visible, but the output, dependent upon the state is visible. The only parameters are state transition probabilities. Each state has a probability distribution over possible o/p tokens.

HMM Markov process is hidden, we can see sequence of output symbols (observations).

HMM - Conditional Dependence

HMM: Simple Example Determine the average annual temperature at a particular location over a series of years (Past when thermometers were not invented). 2 annual temperatures, Hot – H and Cold - C. A correlation between the size of tree growth rings and temperature. We can observe Tree ring size. Temperature is unobserved – hidden.

HMM – Formation of problem 2 hidden states – H and C 3 observed states – tree ring sizes. Small – S, Medium – M, Large – L. The transition probabilities, observation matrix and initial state distribution. All matrices are row stochastic.

HMM – Formation of problem Consider a 4 year sequence. We observe the series of tree rings S;M; S; L. O = (0, 1, 0, 2) We need to determine temperature (H or C) for these 4 years i. e. Most likely state sequence of Markov process given observations.

HMM – Formation of problem X = (x 0, x 1, x 2, x 3 ) O = (O 0, O 1, O 2, O 3 ) A = State transition probability (a ij ) B = Observation probability matrix (b ij )

HMM – Formation of problem a ij = P(state q j at t + 1 | state q i at t) B j (k) = P(observation k at t | state q j at t) P(X) = π x0 * b x0 (O 0 ) * a x0,x1 * b x1 (O 1 ) * a x1,x2 * b x2 (O 2 ) * a x2,x3 b x3 (O 3 ) P(HHCC) = 0.6(0.1)(0.7)(0.4)(0.3)(0.7)(0.6)(0.1) =

Applying HMM to Error Generation Erroneous data in real-world data sets Typing errors are very common. – Insertion – Deletion – Replace Is there any way to determine most probable sequence or patterns of errors made by typist?

Applying HMM to Error Generation Examples: 1. BRIDGETT and BRIDGETTE 2. WILLIAMS and WILIAMS 3. LATONYA and LATOYA 4. FREEMAN and FREEMON

Applying HMM to Error Generation Sequence of characters/Alignment Problem WILLIAMS W I L L L I M

HMM & Error Generation Hidden states: Pointer positions Observations: Output character sequence Problems: o Finding Path - Given an input, output character sequence and HMM model, determine most probable operation sequence? o Training - Given n pairs of input and output sequences, what is the model that maximizes probability of output? o Likelihood - Given input, output and the model, determine likelihood of observed sequence.

References Why should machines learn? – Herbert A. Simon, Department of Computer Science and Psychology, Carnegie-Mellon University, C.I.P. # A Revealing Introduction to Hidden Markov Models – Mark Stamp, Department of Computer Science, San Jose State University

THANK YOU!