Progress Report Reihaneh Rabbany Presented for NLP Group Computing Science Department University of Alberta April 2009.

Slides:



Advertisements
Similar presentations
Cognitive Systems, ICANN panel, Q1 What is machine intelligence, as beyond pattern matching, classification and prediction. What is machine intelligence,
Advertisements

Reinforcement Learning
CS 4100 Artificial Intelligence Prof. C. Hafner Class Notes March 27, 2012.
Eick: Reinforcement Learning. Reinforcement Learning Introduction Passive Reinforcement Learning Temporal Difference Learning Active Reinforcement Learning.
Background Reinforcement Learning (RL) agents learn to do tasks by iteratively performing actions in the world and using resulting experiences to decide.
Patch to the Future: Unsupervised Visual Prediction
Hidden Markov Models Theory By Johan Walters (SR 2003)
GS 540 week 6. HMM basics Given a sequence, and state parameters: – Each possible path through the states has a certain probability of emitting the sequence.
Phylogenetic Trees Presenter: Michael Tung
Huang,Kaizhu Classifier based on mixture of density tree CSE Department, The Chinese University of Hong Kong.
1 Hybrid Agent-Based Modeling: Architectures,Analyses and Applications (Stage One) Li, Hailin.
True music must repeat the thought and inspirations of the people and the time. My people are children and my time is today.
1 Kunstmatige Intelligentie / RuG KI Reinforcement Learning Johan Everts.
Scalable Information-Driven Sensor Querying and Routing for ad hoc Heterogeneous Sensor Networks Maurice Chu, Horst Haussecker and Feng Zhao Xerox Palo.
CPSC 322, Lecture 31Slide 1 Probability and Time: Markov Models Computer Science cpsc322, Lecture 31 (Textbook Chpt 6.5) March, 25, 2009.
Copyright N. Friedman, M. Ninio. I. Pe’er, and T. Pupko. 2001RECOMB, April 2001 Structural EM for Phylogentic Inference Nir Friedman Computer Science &
Learning: Reinforcement Learning Russell and Norvig: ch 21 CMSC421 – Fall 2005.
Learning and Recognizing Activities in Streams of Video Dinesh Govindaraju.
Behaviorist Psychology R+R- P+P- B. F. Skinner’s operant conditioning.
Statistical Natural Language Processing. What is NLP?  Natural Language Processing (NLP), or Computational Linguistics, is concerned with theoretical.
CS 4100 Artificial Intelligence Prof. C. Hafner Class Notes April 3, 2012.
Digital Camera and Computer Vision Laboratory Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan, R.O.C.
Extracting Places and Activities from GPS Traces Using Hierarchical Conditional Random Fields Yong-Joong Kim Dept. of Computer Science Yonsei.
Overview of NIT HMM-based speech synthesis system for Blizzard Challenge 2011 Kei Hashimoto, Shinji Takaki, Keiichiro Oura, and Keiichi Tokuda Nagoya.
Digital Camera and Computer Vision Laboratory Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan, R.O.C.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
EM and expected complete log-likelihood Mixture of Experts
Stochastic Routing Routing Area Meeting IETF 82 (Taipei) Nov.15, 2011.
Machine Learning in Spoken Language Processing Lecture 21 Spoken Language Processing Prof. Andrew Rosenberg.
A brief overview of Speech Recognition and Spoken Language Processing Advanced NLP Guest Lecture August 31 Andrew Rosenberg.
1 ECE-517: Reinforcement Learning in Artificial Intelligence Lecture 6: Optimality Criterion in MDPs Dr. Itamar Arel College of Engineering Department.
Phylogenetic Analysis. General comments on phylogenetics Phylogenetics is the branch of biology that deals with evolutionary relatedness Uses some measure.
Clustering Spatial Data Using Random Walk David Harel and Yehuda Koren KDD 2001.
Modeling Speech using POMDPs In this work we apply a new model, POMPD, in place of the traditional HMM to acoustically model the speech signal. We use.
Comp 538 Course Presentation Discrete Factor Analysis Learning Hidden Variables in Bayesian Network Calvin Hua & Lily Tian Computer Science Dep, HKUST.
LML Speech Recognition Speech Recognition Introduction I E.M. Bakker.
Sequence Models With slides by me, Joshua Goodman, Fei Xia.
Design Principles for Creating Human-Shapable Agents W. Bradley Knox, Ian Fasel, and Peter Stone The University of Texas at Austin Department of Computer.
© D. Weld and D. Fox 1 Reinforcement Learning CSE 473.
L ETTER TO P HONEME A LIGNMENT Reihaneh Rabbany Shahin Jabbari.
Learning the Structure of Related Tasks Presented by Lihan He Machine Learning Reading Group Duke University 02/03/2006 A. Niculescu-Mizil, R. Caruana.
Simultaneously Learning and Filtering Juan F. Mancilla-Caceres CS498EA - Fall 2011 Some slides from Connecting Learning and Logic, Eyal Amir 2006.
Neural Networks Chapter 7
1 CSE 552/652 Hidden Markov Models for Speech Recognition Spring, 2005 Oregon Health & Science University OGI School of Science & Engineering John-Paul.
The famous “sprinkler” example (J. Pearl, Probabilistic Reasoning in Intelligent Systems, 1988)
Digital Camera and Computer Vision Laboratory Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan, R.O.C.
Multi-Speaker Modeling with Shared Prior Distributions and Model Structures for Bayesian Speech Synthesis Kei Hashimoto, Yoshihiko Nankaku, and Keiichi.
CSE 5331/7331 F'07© Prentice Hall1 CSE 5331/7331 Fall 2007 Machine Learning Margaret H. Dunham Department of Computer Science and Engineering Southern.
Eick: Reinforcement Learning. Reinforcement Learning Introduction Passive Reinforcement Learning Temporal Difference Learning Active Reinforcement Learning.
Letter to Phoneme Alignment Using Graphical Models N. Bolandzadeh, R. Rabbany Dept of Computing Science University of Alberta 1 1.
Inferring High-Level Behavior from Low-Level Sensors Donald J. Patterson, Lin Liao, Dieter Fox, and Henry Kautz.
Performance Comparison of Speaker and Emotion Recognition
1 Use graphs and not pure logic Variables represented by nodes and dependencies by edges. Common in our language: “threads of thoughts”, “lines of reasoning”,
Learning and Acting with Bayes Nets Chapter 20.. Page 2 === A Network and a Training Data.
Bayesian Speech Synthesis Framework Integrating Training and Synthesis Processes Kei Hashimoto, Yoshihiko Nankaku, and Keiichi Tokuda Nagoya Institute.
Statistical Models for Automatic Speech Recognition Lukáš Burget.
Reinforcement Learning for Mapping Instructions to Actions S.R.K. Branavan, Harr Chen, Luke S. Zettlemoyer, Regina Barzilay Computer Science and Artificial.
Computational Intelligence: Methods and Applications Lecture 26 Density estimation, Expectation Maximization. Włodzisław Duch Dept. of Informatics, UMK.
Reinforcement Learning Introduction Passive Reinforcement Learning Temporal Difference Learning Active Reinforcement Learning Applications Summary.
Interpreting Ambiguous Emotional Expressions Speech Analysis and Interpretation Laboratory ACII 2009.
Hidden Markov Models Wassnaa AL-mawee Western Michigan University Department of Computer Science CS6800 Adv. Theory of Computation Prof. Elise De Doncker.
Online Multiscale Dynamic Topic Models
Statistical Models for Automatic Speech Recognition
Probability and Time: Markov Models
Announcements Homework 3 due today (grace period through Friday)
Probability and Time: Markov Models
Statistical Models for Automatic Speech Recognition
Probability and Time: Markov Models
Computer Vision Chapter 4
Probability and Time: Markov Models
Presentation transcript:

Progress Report Reihaneh Rabbany Presented for NLP Group Computing Science Department University of Alberta April 2009

Agenda Project Proposal for Guiding Agent by Speech Many to Many Alignment by Bayesian Networks – Letter to Phoneme Alignment – Evaluation of phylogenetic trees 2

Quick RL overview An agent interacting with environment – perceives state – performs actions – receive rewards Agent – Computes the value of each action in each state long term reward obtainable from this state by performing this action – Performs action selection by choosing the best action or sometimes a random action exploration-exploitation 3

Project Proposal for Guiding Agent by Speech Accelerate learning using speech – The emotion in speech signal has considerable amount of side information – Happiness or anger of a speech signal can provide a shaping reinforcement signal Developing tools and methods to extract emotion from speech and designing a methodology to use it as a shaping signal 4

Approaches to use speech signal as a guide for learning Extracting prosodic features from speech Associating meaning to these features – Supervised learning-based approach data-set of (prosodic features, emotion) pairs – excited, happy, upset, sad, bored Assigns a reward to the recognized emotion – Pure RL approach inspired by the learning process of the parent-infant – The infant gradually learns to associate value to perceived speech and how to use it to guide her exploration of the world 5

RL Approach Two ways for developing this idea – Augmenting the observation space to include the prosodic features Emotion will become state-dependent – Learns a separate instructor module Estimates the value of prosodic features Instructions (learnt instructor values) would affect the agent's action selection 6

Instructions Different ways that these instructions (learnt instructor values) could affect the agent's action selection – Balancing the exploration-exploitation When the speaker is not happy with what the agent is doing and it should explore other actions – Use it directly in action selection by some weights Motivates the agent to keep its previous action if the instructor is satisfied with its current action – Use it as a shaping reward to define a new reward function by adding it to the actual reward received from the environment 7

Agenda Project Proposal for Guiding Agent by Speech Many to Many Alignment by Bayesian Networks – Letter to Phoneme Alignment – Evaluation of phylogenetic trees 8

Many to Many Alignment by Bayesian Networks Finding Alignment between two sequences – Assuming the order is preserved I’ve applied it into two applications – Letter to phoneme alignment Aligning for a given dictionary – Evaluating Phylogenetic trees Shows how compatible the tree is with the given taxonomy 9

Agenda Project Proposal for Guiding Agent by Speech Many to Many Alignment by Bayesian Networks – Letter to Phoneme Alignment – Phylogenetic trees evaluation 10

Model Word: – sequence of letters Pronunciation: – sequence of phonemes Alignment: – sequence of subalignments Problem: Finding the most probable alignment Assumption: sub alignments are independent 11

Many-to-Many EM 1. Initialize prob(SubAlignmnets) // Expectation Step 2. For each word in training_set 2.1. Produce all possible alignments 2.2. Choose the most probable alignment // Maximization Step 3. For all subalignments 3.1. Compute new_p(SubAlignmnets) 12

Dynamic Bayesian Network Model Subaligments : hidden variables Learn DBN by EM lili PiPi aiai 13

Context Dependent DBN Context independency assumption – Makes the model simpler – It is not always a correct assumption – Example: P( ) in Chat and Hat Model lili PiPi aiai a i-1 14

Agenda Project Proposal for Guiding Agent by Speech Many to Many Alignment by Bayesian Networks – Letter to Phoneme Alignment – Evaluation of phylogenetic trees 15

Evaluation of Phylogenetic Trees Phylogenetic Trees – Show the evolution of species Taxonomy – Caninae; True dogs; Canis; Coyote – … – Caninae; True foxes; Vulpes; Kit Fox – Caninae; True foxes; Vulpes; Fennec Fox – … – Caninae; Basal Caninae; Otocyon ; Bat-eared Fox –... 16

Tree Evaluation Labeling the inner nodes in the tree For each species – A path in the tree  sequence of inner node labels – A taxonomy description  taxonomy sequence – There should be a many to many alignment between these two sequences 17

Tree Evaluation (Cont.) Finding alignment between these sequences for all the species – Finding the most probable alignments Measuring the mean probability of these alignment – How probable is this tree given this taxonomy 18

Taxonomy and Trees Aligned result 19

Discussion 20