CS 188: Artificial Intelligence Fall 2008

Slides:



Advertisements
Similar presentations
Lecture 16 Hidden Markov Models. HMM Until now we only considered IID data. Some data are of sequential nature, i.e. have correlations have time. Example:
Advertisements

Particle Filtering Sometimes |X| is too big to use exact inference
Building an ASR using HTK CS4706
Lirong Xia Approximate inference: Particle filter Tue, April 1, 2014.
Hidden Markov Models Reading: Russell and Norvig, Chapter 15, Sections
1 Hidden Markov Models (HMMs) Probabilistic Automata Ubiquitous in Speech/Speaker Recognition/Verification Suitable for modelling phenomena which are dynamic.
Lecture 15 Hidden Markov Models Dr. Jianjun Hu mleg.cse.sc.edu/edu/csce833 CSCE833 Machine Learning University of South Carolina Department of Computer.
Advanced Artificial Intelligence
1 Reasoning Under Uncertainty Over Time CS 486/686: Introduction to Artificial Intelligence Fall 2013.
HMMs Tamara Berg CS Artificial Intelligence Many slides throughout the course adapted from Svetlana Lazebnik, Dan Klein, Stuart Russell, Andrew.
Natural Language Processing - Speech Processing -
Application of HMMs: Speech recognition “Noisy channel” model of speech.
… Hidden Markov Models Markov assumption: Transition model:
CS 188: Artificial Intelligence Fall 2009 Lecture 20: Particle Filtering 11/5/2009 Dan Klein – UC Berkeley TexPoint fonts used in EMF. Read the TexPoint.
Hidden Markov Models. Hidden Markov Model In some Markov processes, we may not be able to observe the states directly.
CS 188: Artificial Intelligence Fall 2009 Lecture 21: Speech Recognition 11/10/2009 Dan Klein – UC Berkeley TexPoint fonts used in EMF. Read the TexPoint.
1 Hidden Markov Model Instructor : Saeed Shiry  CHAPTER 13 ETHEM ALPAYDIN © The MIT Press, 2004.
COMP 4060 Natural Language Processing Speech Processing.
Hidden Markov Models. Hidden Markov Model In some Markov processes, we may not be able to observe the states directly.
Dynamic Time Warping Applications and Derivation
CS 188: Artificial Intelligence Fall 2009 Lecture 19: Hidden Markov Models 11/3/2009 Dan Klein – UC Berkeley.
Human Psychoacoustics shows ‘tuning’ for frequencies of speech If a tree falls in the forest and no one is there to hear it, will it make a sound?
Representing Acoustic Information
CHAPTER 15 SECTION 3 – 4 Hidden Markov Models. Terminology.
Audio Processing for Ubiquitous Computing Uichin Lee KAIST KSE.
LE 460 L Acoustics and Experimental Phonetics L-13
Gaussian Mixture Model and the EM algorithm in Speech Recognition
Artificial Intelligence 2004 Speech & Natural Language Processing Natural Language Processing written text as input sentences (well-formed) Speech.
QUIZ!!  T/F: The forward algorithm is really variable elimination, over time. TRUE  T/F: Particle Filtering is really sampling, over time. TRUE  T/F:
Harmonics November 1, 2010 What’s next? We’re halfway through grading the mid-terms. For the next two weeks: more acoustics It’s going to get worse before.
Speech, Perception, & AI Artificial Intelligence CMSC March 5, 2002.
CHAPTER 15 SECTION 1 – 2 Markov Models. Outline Probabilistic Inference Bayes Rule Markov Chains.
Artificial Intelligence 2004 Speech & Natural Language Processing Natural Language Processing written text as input sentences (well-formed) Speech.
UIUC CS 498: Section EA Lecture #21 Reasoning in Artificial Intelligence Professor: Eyal Amir Fall Semester 2011 (Some slides from Kevin Murphy (UBC))
CS 188: Artificial Intelligence Fall 2007 Lecture 22: Viterbi 11/13/2007 Dan Klein – UC Berkeley.
Speech, Perception, & AI Artificial Intelligence CMSC February 13, 2003.
Announcements Project 4: Ghostbusters Homework 7
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
Hidden Markov Models & POS Tagging Corpora and Statistical Methods Lecture 9.
Statistical NLP Spring 2011
CHAPTER 8 DISCRIMINATIVE CLASSIFIERS HIDDEN MARKOV MODELS.
The famous “sprinkler” example (J. Pearl, Probabilistic Reasoning in Intelligent Systems, 1988)
Hidden Markovian Model. Some Definitions Finite automation is defined by a set of states, and a set of transitions between states that are taken based.
QUIZ!!  In HMMs...  T/F:... the emissions are hidden. FALSE  T/F:... observations are independent given no evidence. FALSE  T/F:... each variable X.
Probabilistic reasoning over time Ch. 15, 17. Probabilistic reasoning over time So far, we’ve mostly dealt with episodic environments –Exceptions: games.
Hidden Markov Models (HMMs) –probabilistic models for learning patterns in sequences (e.g. DNA, speech, weather, cards...) (2 nd order model)
CS 188: Artificial Intelligence Spring 2007 Speech Recognition 03/20/2007 Srini Narayanan – ICSI and UC Berkeley.
Reasoning over Time  Often, we want to reason about a sequence of observations  Speech recognition  Robot localization  User attention  Medical monitoring.
1 Hidden Markov Model Xiaole Shirley Liu STAT115, STAT215.
Harmonics October 28, Where Were We? Mid-terms: our goal is to get them back to you by Friday. Production Exercise #2 results should be sent to.
Automatic Speech Recognition
The Physics of Sound.
HMM: Particle filters Lirong Xia. HMM: Particle filters Lirong Xia.
Lecture 9: Speech Recognition (I) October 26, 2004 Dan Jurafsky
Speech Processing Speech Recognition
CS 188: Artificial Intelligence Spring 2007
CAP 5636 – Advanced Artificial Intelligence
Sound shadow effect Depends on the size of the obstructing object and the wavelength of the sound. If comparable: Then sound shadow occurs. I:\users\mnshriv\3032.
CS 188: Artificial Intelligence
Hidden Markov Models Markov chains not so useful for most agents
CS 188: Artificial Intelligence Spring 2006
Algorithms of POS Tagging
LECTURE 15: REESTIMATION, EM AND MIXTURES
Speech recognition, machine learning
Speech Recognition: Acoustic Waves
CS 188: Artificial Intelligence Spring 2006
HMM: Particle filters Lirong Xia. HMM: Particle filters Lirong Xia.
CS 188: Artificial Intelligence Fall 2008
Speech recognition, machine learning
Presentation transcript:

CS 188: Artificial Intelligence Fall 2008 Lecture 21: Speech / Viterbi 11/13/2008 Dan Klein – UC Berkeley

Announcements P5 up, due 11/19 W9 up, due 11/21 (note off-cycle date) Final contest: download and get started! Homework solution and review sessions coming

Hidden Markov Models An HMM is Query: most likely seq: X1 X2 X3 X4 X5 E1 E2 E3 E4 E5 An HMM is Initial distribution: Transitions: Emissions: Query: most likely seq:

State Path Trellis State trellis: graph of states and transitions over time Each arc represents some transition Each arc has weight Each path is a sequence of states The product of weights on a path is the seq’s probability Can think of the Forward (and now Viterbi) algorithms as computing sums of all paths (best paths) in this graph sun sun sun sun rain rain rain rain

Viterbi Algorithm sun sun sun sun rain rain rain rain

Example

Digitizing Speech

Speech in an Hour Speech input is an acoustic wave form s p ee ch l a b “l” to “a” transition: Graphs from Simon Arnfield’s web tutorial on speech, Sheffield: http://www.psyc.leeds.ac.uk/research/cogn/speech/tutorial/

Spectral Analysis Frequency gives pitch; amplitude gives volume sampling at ~8 kHz phone, ~16 kHz mic (kHz=1000 cycles/sec) Fourier transform of wave displayed as a spectrogram darkness indicates energy at each frequency s p ee ch l a b amplitude frequency

Adding 100 Hz + 1000 Hz Waves

Spectrum Frequency components (100 and 1000 Hz) on x-axis Amplitude Frequency in Hz

Part of [ae] from “lab” Note complex wave repeating nine times in figure Plus smaller waves which repeats 4 times for every large pattern Large wave has frequency of 250 Hz (9 times in .036 seconds) Small wave roughly 4 times this, or roughly 1000 Hz Two little tiny waves on top of peak of 1000 Hz waves

Back to Spectra Spectrum represents these freq components Computed by Fourier transform, algorithm which separates out each frequency component of wave. x-axis shows frequency, y-axis shows magnitude (in decibels, a log measure of amplitude) Peaks at 930 Hz, 1860 Hz, and 3020 Hz.

Acoustic Feature Sequence Time slices are translated into acoustic feature vectors (~39 real numbers per slice) These are the observations, now we need the hidden states X frequency ……………………………………………..e12e13e14e15e16………..

State Space P(E|X) encodes which acoustic vectors are appropriate for each phoneme (each kind of sound) P(X|X’) encodes how sounds can be strung together We will have one state for each sound in each word From some state x, can only: Stay in the same state (e.g. speaking slowly) Move to the next position in the word At the end of the word, move to the start of the next word We build a little state graph for each word and chain them together to form our state space X

HMMs for Speech

Markov Process with Bigrams Figure from Huang et al page 618

Decoding While there are some practical issues, finding the words given the acoustics is an HMM inference problem We want to know which state sequence x1:T is most likely given the evidence e1:T: From the sequence x, we can simply read off the words

End of Part II! Now we’re done with our unit on probabilistic reasoning Last part of class: machine learning