Pattern Finding and Pattern Discovery in Time Series

Slides:



Advertisements
Similar presentations
1 Gesture recognition Using HMMs and size functions.
Advertisements

FEATURE PERFORMANCE COMPARISON FEATURE PERFORMANCE COMPARISON y SC is a training set of k-dimensional observations with labels S and C b C is a parameter.
. Lecture #8: - Parameter Estimation for HMM with Hidden States: the Baum Welch Training - Viterbi Training - Extensions of HMM Background Readings: Chapters.
Lecture 16 Hidden Markov Models. HMM Until now we only considered IID data. Some data are of sequential nature, i.e. have correlations have time. Example:
Change Detection C. Stauffer and W.E.L. Grimson, “Learning patterns of activity using real time tracking,” IEEE Trans. On PAMI, 22(8): , Aug 2000.
Angelo Dalli Department of Intelligent Computing Systems
HMM II: Parameter Estimation. Reminder: Hidden Markov Model Markov Chain transition probabilities: p(S i+1 = t|S i = s) = a st Emission probabilities:
Learning HMM parameters
Hidden Markov Models. A Hidden Markov Model consists of 1.A sequence of states {X t |t  T } = {X 1, X 2,..., X T }, and 2.A sequence of observations.
Hidden Markov models and its application to bioinformatics.
Tutorial on Hidden Markov Models.
2004/11/161 A Tutorial on Hidden Markov Models and Selected Applications in Speech Recognition LAWRENCE R. RABINER, FELLOW, IEEE Presented by: Chi-Chun.
Patterns, Profiles, and Multiple Alignment.
Ch 9. Markov Models 고려대학교 자연어처리연구실 한 경 수
Statistical NLP: Lecture 11
Hidden Markov Models Theory By Johan Walters (SR 2003)
Foundations of Statistical NLP Chapter 9. Markov Models 한 기 덕한 기 덕.
1 Hidden Markov Models (HMMs) Probabilistic Automata Ubiquitous in Speech/Speaker Recognition/Verification Suitable for modelling phenomena which are dynamic.
Hidden Markov Models Fundamentals and applications to bioinformatics.
PatReco: Hidden Markov Models Alexandros Potamianos Dept of ECE, Tech. Univ. of Crete Fall
HMM-BASED PATTERN DETECTION. Outline  Markov Process  Hidden Markov Models Elements Basic Problems Evaluation Optimization Training Implementation 2-D.
Hidden Markov Model 11/28/07. Bayes Rule The posterior distribution Select k with the largest posterior distribution. Minimizes the average misclassification.
Hidden Markov Models I Biology 162 Computational Genetics Todd Vision 14 Sep 2004.
Lecture 5: Learning models using EM
Hidden Markov Models K 1 … 2. Outline Hidden Markov Models – Formalism The Three Basic Problems of HMMs Solutions Applications of HMMs for Automatic Speech.
. Hidden Markov Models with slides from Lise Getoor, Sebastian Thrun, William Cohen, and Yair Weiss.
Hidden Markov Models David Meir Blei November 1, 1999.
Learning HMM parameters Sushmita Roy BMI/CS 576 Oct 21 st, 2014.
. cmsc726: HMMs material from: slides from Sebastian Thrun, and Yair Weiss.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Advanced Signal Processing 2, SE Professor Horst Cerjak, Andrea Sereinig Graz, Basics of Hidden Markov Models Basics of HMM-based.
Isolated-Word Speech Recognition Using Hidden Markov Models
Segmental Hidden Markov Models with Random Effects for Waveform Modeling Author: Seyoung Kim & Padhraic Smyth Presentor: Lu Ren.
Fundamentals of Hidden Markov Model Mehmet Yunus Dönmez.
Bayesian networks Classification, segmentation, time series prediction and more. Website: Twitter:
Hidden Markov Models Usman Roshan CS 675 Machine Learning.
Cognitive Computer Vision Kingsley Sage and Hilary Buxton Prepared under ECVision Specific Action 8-3
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: ML and Simple Regression Bias of the ML Estimate Variance of the ML Estimate.
S. Salzberg CMSC 828N 1 Three classic HMM problems 2.Decoding: given a model and an output sequence, what is the most likely state sequence through the.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
CS Statistical Machine learning Lecture 24
1 CONTEXT DEPENDENT CLASSIFICATION  Remember: Bayes rule  Here: The class to which a feature vector belongs depends on:  Its own value  The values.
CHAPTER 8 DISCRIMINATIVE CLASSIFIERS HIDDEN MARKOV MODELS.
Computer Vision Lecture 6. Probabilistic Methods in Segmentation.
1 CS 552/652 Speech Recognition with Hidden Markov Models Winter 2011 Oregon Health & Science University Center for Spoken Language Understanding John-Paul.
1 CSE 552/652 Hidden Markov Models for Speech Recognition Spring, 2006 Oregon Health & Science University OGI School of Science & Engineering John-Paul.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Elements of a Discrete Model Evaluation.
Maximum Entropy Model, Bayesian Networks, HMM, Markov Random Fields, (Hidden/Segmental) Conditional Random Fields.
Hidden Markov Models (HMMs) –probabilistic models for learning patterns in sequences (e.g. DNA, speech, weather, cards...) (2 nd order model)
1 Hidden Markov Models Hsin-min Wang References: 1.L. R. Rabiner and B. H. Juang, (1993) Fundamentals of Speech Recognition, Chapter.
Statistical Models for Automatic Speech Recognition Lukáš Burget.
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
DISCRETE HIDDEN MARKOV MODEL IMPLEMENTATION DIGITAL SPEECH PROCESSING HOMEWORK #1 DISCRETE HIDDEN MARKOV MODEL IMPLEMENTATION Date: Oct, Revised.
Automated Speach Recognotion Automated Speach Recognition By: Amichai Painsky.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
ECE 8443 – Pattern Recognition Objectives: Reestimation Equations Continuous Distributions Gaussian Mixture Models EM Derivation of Reestimation Resources:
Hidden Markov Model Parameter Estimation BMI/CS 576 Colin Dewey Fall 2015.
Hidden Markov Models. A Hidden Markov Model consists of 1.A sequence of states {X t |t  T } = {X 1, X 2,..., X T }, and 2.A sequence of observations.
Definition of the Hidden Markov Model A Seminar Speech Recognition presentation A Seminar Speech Recognition presentation October 24 th 2002 Pieter Bas.
Other Models for Time Series. The Hidden Markov Model (HMM)
Visual Recognition Tutorial1 Markov models Hidden Markov models Forward/Backward algorithm Viterbi algorithm Baum-Welch estimation algorithm Hidden.
Date: October, Revised by 李致緯
An Iterative Approach to Discriminative Structure Learning
Hidden Markov Models Part 2: Algorithms
Hidden Markov Model LR Rabiner
CONTEXT DEPENDENT CLASSIFICATION
LECTURE 15: REESTIMATION, EM AND MIXTURES
Introduction to HMM (cont)
Presentation transcript:

Pattern Finding and Pattern Discovery in Time Series Long Q Tran College of Computing, Georgia Tech Trần Quốc Long College of Computing, Georgia Tech

Contents Pattern Finding & Pattern Discovery Pattern Finding & Pattern Discovery in Time Series Hidden Markov Models (HMMs) Summary

Pattern Finding Problems: given observed patterns O1, O2, … OK, specify which pattern the new data X possess? Other names: pattern recognition, pattern classification Examples Recognition: matching fingerprints of the claimant with those of authorized personnel.

Pattern Finding Patterns are known beforehand and are observed/described by Explicit samples Similar samples (usually) Modeling approaches: Build a model for each pattern Find the best fit model for new data Usually require training using observed samples

Pattern Discovery Patterns are not known But data which are believed to possess patterns are given Examples: Clustering: grouping similar samples into clusters Associative rule mining: discover certain features that often appear together in data

Contents Pattern Finding & Pattern Discovery Pattern Finding & Pattern Discovery in Time Series Hidden Markov Models (HMMs) Summary

Time Series Data are sampled over time X = X1 X2 … Xt … XL Xt : data sampled at time t L : sequence length Xt are NOT independently and identically distributed (NOT i.d.d) In other words, Xt may come from different processes that are dependent of each other

Pattern Finding in Time Series Examples In control, certain pattern of sensor signals indicate critical point of the production process In stock, certain pattern (up/down) of price indicate the trend of the market People often have to look at the graph by their own eyes and act accordingly when spotting known pattern X. Ge & P. Smyth (2000): detecting end-point in plasma etch (semiconductor manufacturing)

Pattern Finding in Time Series Problems: Data may contain one or more patterns inside Data can be multi-dimensional (i.e. look at multiple graphs at the same time) Automated pattern finding is crucial when time series are lengthy and multi-dimensional

Pattern Discovery in Time Series Goals: From collected data, discover Replicated, interesting patterns Associative rule on patterns (can use to predict trends of time series)

Pattern Modeling in Time Series Both pattern finding and pattern discovery need modeling Desired properties of the model The model can be built or trained using observed data The similarity of new data and the model can be easily computed

Contents Pattern Finding & Pattern Discovery Pattern Finding & Pattern Discovery in Time Series Hidden Markov Models (HMMs) Summary

Hidden Markov Models (HMMs) One way to model time series pattern Assumptions: Xt is generated from certain probability distribution Yt (called state) Number of states is finite (i.e. finite sources of data) State transition follows Markov property X1 Y1 X2 Y2 XL YL … 1 2 0.6 0.4 1 2 0.4 0.6

Hidden Markov Models (HMMs) Parameters to estimate: Transition probabilities Distribution parameters in each state Estimation procedure: Initialization: k-means, viterbi training Iterative training: forward-backward procedure (EM algorithm) Variants of HMM: Mixture of HMMs: allow many HMMs computed simultaneously State durational HMM: allow a state remains for a duration

Mixture of HMMs Assumption: Mixture of HMMs allows There are different processes (pattern) that generate the time series Each process can be represented by a HMM Mixture of HMMs allows Packing all pattern models in one place Identifying the processes that generate the time series Training be efficiently implemented

Experiment Experiment settings Generate 200 sequences for each HMM 1 2 0.6 0.4 1 2 0.4 0.6 Experiment settings Generate 200 sequences for each HMM After 200 iterations 2 Gaussian:  = 0,  = 1 1 Gaussian:  = 2,  = 1 2 11 = -0.07, 11 = 0.97 21 = 2.01, 21 = 0.99 12 = 1.90, 12 = 1.10 22 = -0.01, 22 = 0.98

Summary Automated pattern finding and pattern discovery in time series are needed HMMs and its variants can model time series patterns Parameters can be efficiently initialized and estimated using observed data

 = (transition prob., distribution params.) Appendix: HMMs Parameters:  = (transition prob., distribution params.) Recognition Calculate P(X1X2…XL |) Forward procedure Estimation: Maximize L() = P(X1X2…XL |) EM algorithm: forward – backward procedure Clustering Find Viterbi algorithm