22CS 338: Graphical User Interfaces. Dario Salvucci, Drexel University. Lecture 10: Advanced Input.

Slides:



Advertisements
Similar presentations
1 Gesture recognition Using HMMs and size functions.
Advertisements

Lecture 16 Hidden Markov Models. HMM Until now we only considered IID data. Some data are of sequential nature, i.e. have correlations have time. Example:
Character Recognition using Hidden Markov Models Anthony DiPirro Ji Mei Sponsor:Prof. William Sverdlik.
Speech Recognition Part 3 Back end processing. Speech recognition simplified block diagram Speech Capture Speech Capture Feature Extraction Feature Extraction.
Learning HMM parameters
Introduction to Hidden Markov Models
Hidden Markov Models Bonnie Dorr Christof Monz CMSC 723: Introduction to Computational Linguistics Lecture 5 October 6, 2004.
O PTICAL C HARACTER R ECOGNITION USING H IDDEN M ARKOV M ODELS Jan Rupnik.
An Introduction to Hidden Markov Models and Gesture Recognition Troy L. McDaniel Research Assistant Center for Cognitive Ubiquitous Computing Arizona State.
Hidden Markov Models Adapted from Dr Catherine Sweeney-Reed’s slides.
Ch-9: Markov Models Prepared by Qaiser Abbas ( )
Hidden Markov Models Theory By Johan Walters (SR 2003)
1 Hidden Markov Models (HMMs) Probabilistic Automata Ubiquitous in Speech/Speaker Recognition/Verification Suitable for modelling phenomena which are dynamic.
Hidden Markov Models in NLP
Lecture 15 Hidden Markov Models Dr. Jianjun Hu mleg.cse.sc.edu/edu/csce833 CSCE833 Machine Learning University of South Carolina Department of Computer.
Hidden Markov Models 1 2 K … 1 2 K … 1 2 K … … … … 1 2 K … x1x1 x2x2 x3x3 xKxK 2 1 K 2.
SPEECH RECOGNITION Kunal Shalia and Dima Smirnov.
INTRODUCTION TO Machine Learning 3rd Edition
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
HMM-BASED PATTERN DETECTION. Outline  Markov Process  Hidden Markov Models Elements Basic Problems Evaluation Optimization Training Implementation 2-D.
Hidden Markov Models. Hidden Markov Model In some Markov processes, we may not be able to observe the states directly.
Hidden Markov Models K 1 … 2. Outline Hidden Markov Models – Formalism The Three Basic Problems of HMMs Solutions Applications of HMMs for Automatic Speech.
1 Hidden Markov Model Instructor : Saeed Shiry  CHAPTER 13 ETHEM ALPAYDIN © The MIT Press, 2004.
Doug Downey, adapted from Bryan Pardo,Northwestern University
Learning HMM parameters Sushmita Roy BMI/CS 576 Oct 21 st, 2014.
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Visual Recognition Tutorial1 Markov models Hidden Markov models Forward/Backward algorithm Viterbi algorithm Baum-Welch estimation algorithm Hidden.
Handwritten Character Recognition using Hidden Markov Models Quantifying the marginal benefit of exploiting correlations between adjacent characters and.
Introduction to Automatic Speech Recognition
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Combined Lecture CS621: Artificial Intelligence (lecture 25) CS626/449: Speech-NLP-Web/Topics-in- AI (lecture 26) Pushpak Bhattacharyya Computer Science.
Isolated-Word Speech Recognition Using Hidden Markov Models
1 7-Speech Recognition (Cont’d) HMM Calculating Approaches Neural Components Three Basic HMM Problems Viterbi Algorithm State Duration Modeling Training.
Ondrej Rohlik, Pavel Mautner, Vaclav Matousek, Juergen Kempf
11.10 Human Computer Interface www. ICT-Teacher.com.
7-Speech Recognition Speech Recognition Concepts
Fundamentals of Hidden Markov Model Mehmet Yunus Dönmez.
A brief overview of Speech Recognition and Spoken Language Processing Advanced NLP Guest Lecture August 31 Andrew Rosenberg.
Csc Lecture 7 Recognizing speech. Geoffrey Hinton.
Online Arabic Handwriting Recognition Fadi Biadsy Jihad El-Sana Nizar Habash Abdul-Rahman Daud Done byPresented by KFUPM Information & Computer Science.
Dialog Design - Gesture & Pen Interfaces, Mobile Devices CS / Psych This material has been developed by Georgia Tech HCI faculty, and continues.
Fall 2002CS/PSY Dialog Design 3 How to use a PDA.
Sequence Models With slides by me, Joshua Goodman, Fei Xia.
Automatic Speech Recognition: Conditional Random Fields for ASR Jeremy Morris Eric Fosler-Lussier Ray Slyh 9/19/2008.
Speech, Perception, & AI Artificial Intelligence CMSC February 13, 2003.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
PGM 2003/04 Tirgul 2 Hidden Markov Models. Introduction Hidden Markov Models (HMM) are one of the most common form of probabilistic graphical models,
CS Statistical Machine learning Lecture 24
Probabilistic reasoning over time Ch. 15, 17. Probabilistic reasoning over time So far, we’ve mostly dealt with episodic environments –Exceptions: games.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Elements of a Discrete Model Evaluation.
Maximum Entropy Model, Bayesian Networks, HMM, Markov Random Fields, (Hidden/Segmental) Conditional Random Fields.
Hidden Markov Models (HMMs) –probabilistic models for learning patterns in sequences (e.g. DNA, speech, weather, cards...) (2 nd order model)
Statistical Models for Automatic Speech Recognition Lukáš Burget.
EEL 6586: AUTOMATIC SPEECH PROCESSING Hidden Markov Model Lecture Mark D. Skowronski Computational Neuro-Engineering Lab University of Florida March 31,
Automated Speach Recognotion Automated Speach Recognition By: Amichai Painsky.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
Hidden Markov Model Parameter Estimation BMI/CS 576 Colin Dewey Fall 2015.
Hidden Markov Models. A Hidden Markov Model consists of 1.A sequence of states {X t |t  T } = {X 1, X 2,..., X T }, and 2.A sequence of observations.
Definition of the Hidden Markov Model A Seminar Speech Recognition presentation A Seminar Speech Recognition presentation October 24 th 2002 Pieter Bas.
Other Models for Time Series. The Hidden Markov Model (HMM)
1 7-Speech Recognition Speech Recognition Concepts Speech Recognition Approaches Recognition Theories Bayse Rule Simple Language Model P(A|W) Network Types.
Visual Recognition Tutorial1 Markov models Hidden Markov models Forward/Backward algorithm Viterbi algorithm Baum-Welch estimation algorithm Hidden.
Hidden Markov Models HMM Hassanin M. Al-Barhamtoshy
Dialog Design 3 How to use a PDA
Statistical Models for Automatic Speech Recognition
An INTRODUCTION TO HIDDEN MARKOV MODEL
Statistical Models for Automatic Speech Recognition
Handwritten Characters Recognition Based on an HMM Model
LECTURE 15: REESTIMATION, EM AND MIXTURES
Presentation transcript:

22CS 338: Graphical User Interfaces. Dario Salvucci, Drexel University. Lecture 10: Advanced Input

CS 338: Graphical User Interfaces. Dario Salvucci, Drexel University.22 GUI Input Modalities How can a user provide input to the GUI? – keyboard – mouse, touch pad – joystick – touch screen – pen / stylus – speech – other… more error! harder! (?)

CS 338: Graphical User Interfaces. Dario Salvucci, Drexel University.22 Pen input Pen-based interfaces in mobile computing

CS 338: Graphical User Interfaces. Dario Salvucci, Drexel University.22 Pen input Handwriting – very general, well-developed human skill thus, make use of what users can already do! – but hard to recognize (for people & machines) Gestures – gesture alphabets Palm Pilot graffiti – editing gestures – easier to recognize

CS 338: Graphical User Interfaces. Dario Salvucci, Drexel University.22 Speech input Limited speech recognition – only allow small sets of words/phrases e.g., “one” – “nine” for phone menus Full speech recognition – again, general, well-developed skill – full standard vocabulary ( words) American English: 600,000 words. – specialized vocabularies (research, medical, …) “editing” vocabularies (back, delete, …)

CS 338: Graphical User Interfaces. Dario Salvucci, Drexel University.22 Handwriting & Speech Common issues – vocabulary size – individual variability speaker dependent, adaptive, independent – signal segmentation isolated words, continuous Let’s look at handwriting as an example, but almost all concepts apply to speech too – not to mention other inputs, such as eye movements!

CS 338: Graphical User Interfaces. Dario Salvucci, Drexel University.22 Off-line recognition – examine static output of handwriting, i.e., the end result of the writing On-line recognition – examine dynamic movement of handwriting, i.e., the strokes, pen up/downs involved Which is more “informed”? more useful? Off-line vs. on-line recognition

CS 338: Graphical User Interfaces. Dario Salvucci, Drexel University.22 Recognition techniques Neural networks – neurally-inspired computational models – input: bitmap, or “vectorized” strokes – output: probably characters – best for off-line recognition Hidden Markov models (HMMs) – powerful probabilistic models – input: vectorized strokes – output: full recognition of chars, words, etc. – best for on-line recognition Consider HMM-based on-line recognition… And hybrid methods!

CS 338: Graphical User Interfaces. Dario Salvucci, Drexel University.22 On-line feature extraction On-line strokes  feature vectors – basic features: pen up/down, direction, velocity – useful features: curvature, reversal,...

CS 338: Graphical User Interfaces. Dario Salvucci, Drexel University.22 On-line recognition Hidden Markov models (HMMs) – probabilistic models for dynamic behavior Set of N states with – a(i,j) = probability of state transition i  j – b(o,i) = probability of seeing o in state i can be discrete or continuous prob. distributions s1 b(o,1) s2 b(o,2) s3 b(o,3) a(1,1)a(2,2)a(3,3) a(1,2)a(2,3)

CS 338: Graphical User Interfaces. Dario Salvucci, Drexel University.22 Hidden Markov models Let’s say we have – M = HMM representing predicted behavior – O = observation vector sequence O Three problems – evaluation: find Pr(O|M) – decoding: find the state sequence Q that maximizes Pr(O|M,Q) – training: adjust parameters of M to increase Pr(O|M)

CS 338: Graphical User Interfaces. Dario Salvucci, Drexel University.22 Hidden Markov models HMM evaluation – find Pr(O|M) – evaluate O = … – can we do this efficiently? x =.8 y =.2 x =.5 y =.5 x =.2 y =.8 x =.6 y =

CS 338: Graphical User Interfaces. Dario Salvucci, Drexel University.22 Hidden Markov models HMM decoding (Viterbi algorithm) – find best state sequence through HMM, maximizing the probability of the sequence – Given the word “Hello” Maximize the possible paths through the model by calculating their probability of being actual words. Its less probable that I meant: “Hallo” or “Hejjo” It is more proabable that I meant: – Case 1, word matching:“Jello” – Case 2, letter matching (vector strokes): “HeIIo”

CS 338: Graphical User Interfaces. Dario Salvucci, Drexel University.22 Hidden Markov models HMM training (Baum-Welch / EM algorithm) – re-adjust a(i,j), b(o,i) to increase Pr(O|M) – iterative procedure – allows for fine-tuning of HMM parameters for particular observation sets Every time I write, “HeIIo” I really mean “Hello”. – (Increase in the Vector Stroke probability) – susceptible to just my behaviour. The machine learns the probability of my “Hello”, not your “Hello”.

CS 338: Graphical User Interfaces. Dario Salvucci, Drexel University.22 Hidden Markov models Composing HMMs – we can add “sub-” HMMs into larger HMMs, creating a model hierarchy at different levels For instance, we can create three levels – strokes – letters – words

CS 338: Graphical User Interfaces. Dario Salvucci, Drexel University.22 On-line recognition Stroke HMMs with states – up-down loop s1: up, + curvature, hi velocity s2: down, + curvature, hi velocity – up-down cusp s1: up, + curvature,, hi velocity s2: 0 velocity s3: down, + curvature, hi velocity – up-down ramphoid s1: up, – curvature, hi velocity s2: 0 velocity s3: down, + curvature, hi velocity

CS 338: Graphical User Interfaces. Dario Salvucci, Drexel University.22 On-line recognition Letter HMMs based on stroke HMMs ramphoidcusp loopdown-up loop (‘o’ ?) (‘w’ ?) (‘g’ ?)

CS 338: Graphical User Interfaces. Dario Salvucci, Drexel University.22 On-line recognition Word HMMs based on letter HMMs – basic idea is straightforward – but it’s deceptively tricky — why?? HELLOPLATYPUS

CS 338: Graphical User Interfaces. Dario Salvucci, Drexel University.22 On-line recognition Putting it all together – compacting states – taking word frequencies into account – where do frequencies come from? :-) Note: Probabilities aren’t real… H E L….07 I … A …

CS 338: Graphical User Interfaces. Dario Salvucci, Drexel University.22 Handwriting problems Speed-accuracy tradeoff – as people speed up, their handwriting degrades (uh, no duh!) Printed vs. handwritten? – often some combination of the two!! Dotting i’s, crossing t’s, … for on-line recog’n (minding your p’s and q’s?) Mixing language with graphics & gestures

CS 338: Graphical User Interfaces. Dario Salvucci, Drexel University.22 Speech recognition Same basic ideas for recognition – convert to recognizable signal (transforms) – recognize using hybrid methods and a hierarchy of phonemes, words, etc. Many similar / analogous problems – individual variability (esp. female/male voices) – mixing “real input” with command input – speed-accuracy tradeoff (?)

CS 338: Graphical User Interfaces. Dario Salvucci, Drexel University.22 Eye-movement recognition (?!) – Yet again, same idea: translate noisy signal to what people actually intended – Example: “Eye-typing” system

CS 338: Graphical User Interfaces. Dario Salvucci, Drexel University.22 Discussion Have you used handwriting/speech systems? What are the benefits of these systems? What is handwriting/speech good for? When is it easier to use standard input?