Data Driven Response Generation in Social Media

Slides:



Advertisements
Similar presentations
How do you study for a test ?
Advertisements

Listening skills: how to untangle the noise and find the message Judy Copage.
Statistical Machine Translation
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
The Application of Machine Translation in CADAL Huang Chen, Chen Haiying Zhejiang University Libraries, Hangzhou, China
The Chinese Room: Understanding and Correcting Machine Translation This work has been supported by NSF Grants IIS Solution: The Chinese Room Conclusions.
Statistical Machine Translation Part II – Word Alignments and EM Alex Fraser Institute for Natural Language Processing University of Stuttgart
Proceedings of the Conference on Intelligent Text Processing and Computational Linguistics (CICLing-2007) Learning for Semantic Parsing Advisor: Hsin-His.
Unsupervised Modeling of Twitter Conversations
SEARCHING QUESTION AND ANSWER ARCHIVES Dr. Jiwoon Jeon Presented by CHARANYA VENKATESH KUMAR.
Word Sense Disambiguation for Machine Translation Han-Bin Chen
For Friday No reading Homework –Chapter 23, exercises 1, 13, 14, 19 –Not as bad as it sounds –Do them IN ORDER – do not read ahead here.
Vamshi Ambati | Stephan Vogel | Jaime Carbonell Language Technologies Institute Carnegie Mellon University A ctive Learning and C rowd-Sourcing for Machine.
Confidence Estimation for Machine Translation J. Blatz et.al, Coling 04 SSLI MTRG 11/17/2004 Takahiro Shinozaki.
Jeff Yan School of Computing Science Newcastle University, UK (Joint work with Ahmad Salah El Ahmad) Usability of CAPTCHAs Or “usability issues in CAPTCHA.
Transient Life: Collecting and sharing personal information Stephanie Smale and Saul Greenberg University of Calgary, Canada.
Natural Language Query Interface Mostafa Karkache & Bryce Wenninger.
ITCS 6010 Natural Language Understanding. Natural Language Processing What is it? Studies the problems inherent in the processing and manipulation of.
Extracting Interest Tags from Twitter User Biographies Ying Ding, Jing Jiang School of Information Systems Singapore Management University AIRS 2014, Kuching,
1 The Web as a Parallel Corpus  Parallel corpora are useful  Training data for statistical MT  Lexical correspondences for cross-lingual IR  Early.
1 Statistical NLP: Lecture 13 Statistical Alignment and Machine Translation.
Microsoft Research Faculty Summit Robert Moore Principal Researcher Microsoft Research.
Challenges in Information Retrieval and Language Modeling Michael Shepherd Dalhousie University Halifax, NS Canada.
Finding parallel texts on the web using cross-language information retrieval Achim Ruopp Joint work with Fei Xia University of Washington.
English-Persian SMT Reza Saeedi 1 WTLAB Wednesday, May 25, 2011.
Query Rewriting Using Monolingual Statistical Machine Translation Stefan Riezler Yi Liu Google 2010 Association for Computational Linguistics.
The Five Step Sales Process The Five Step Sales Process Step One: Plan and Prepare May 11, 2011.
Stochastic k-Neighborhood Selection for Supervised and Unsupervised Learning University of Toronto Machine Learning Seminar Feb 21, 2013 Kevin SwerskyIlya.
An Integrated Approach for Arabic-English Named Entity Translation Hany Hassan IBM Cairo Technology Development Center Jeffrey Sorensen IBM T.J. Watson.
Advanced Signal Processing 05/06 Reinisch Bernhard Statistical Machine Translation Phrase Based Model.
1 Computational Linguistics Ling 200 Spring 2006.
Mining the Web to Create Minority Language Corpora Rayid Ghani Accenture Technology Labs - Research Rosie Jones Carnegie Mellon University Dunja Mladenic.
Accent Colors Accent 1 (RGB: 141,178,24) Accent 2 (RGB: 63,63,63) Accent 3 (RGB: 127,127,127) Accent 4 (RGB: 195,214,155) Accent 5 (RGB: 222,222,222) FLAVIUS.
Automatic Post-editing (pilot) Task Rajen Chatterjee, Matteo Negri and Marco Turchi Fondazione Bruno Kessler [ chatterjee | negri | turchi
Why Not Grab a Free Lunch? Mining Large Corpora for Parallel Sentences to Improve Translation Modeling Ferhan Ture and Jimmy Lin University of Maryland,
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
Introduction to Dialogue Systems. User Input System Output ?
Improving Search Results Quality by Customizing Summary Lengths Michael Kaisser ★, Marti Hearst  and John B. Lowe ★ University of Edinburgh,  UC Berkeley,
For Monday Read chapter 24, sections 1-3 Homework: –Chapter 23, exercise 8.
For Friday Finish chapter 24 No written homework.
For Monday Read chapter 26 Last Homework –Chapter 23, exercise 7.
FEISGILTT Dublin 2014 Yves Savourel ENLASO Corporation QuEst Integration in Okapi This presentation was made possible by This project is sponsored by the.
ACE TESOL Diploma Program – London Language Institute OBJECTIVES You will understand: 1. The terminology and concepts of semantics, pragmatics and discourse.
By Jonathan Drake.  The Gnutella protocol is simply not scalable  This is due to the flooding approach it currently utilizes  As the nodes increase.
LREC 2008 Marrakech 29 May Caroline Lavecchia, Kamel Smaïli and David Langlois LORIA / Groupe Parole, Vandoeuvre-Lès-Nancy, France Phrase-Based Machine.
August 17, 2005Question Answering Passage Retrieval Using Dependency Parsing 1/28 Question Answering Passage Retrieval Using Dependency Parsing Hang Cui.
For Friday Finish chapter 23 Homework –Chapter 23, exercise 15.
A New Approach for English- Chinese Named Entity Alignment Donghui Feng Yayuan Lv Ming Zhou USC MSR Asia EMNLP-04.
Statistical Machine Translation Part II: Word Alignments and EM Alex Fraser Institute for Natural Language Processing University of Stuttgart
Discriminative Modeling extraction Sets for Machine Translation Author John DeNero and Dan KleinUC Berkeley Presenter Justin Chiu.
A Framework to Predict the Quality of Answers with Non-Textual Features Jiwoon Jeon, W. Bruce Croft(University of Massachusetts-Amherst) Joon Ho Lee (Soongsil.
For Monday Read chapter 26 Homework: –Chapter 23, exercises 8 and 9.
Question Answering Passage Retrieval Using Dependency Relations (SIGIR 2005) (National University of Singapore) Hang Cui, Renxu Sun, Keya Li, Min-Yen Kan,
Predicting and Adapting to Poor Speech Recognition in a Spoken Dialogue System Diane J. Litman AT&T Labs -- Research
Harnessing the Deep Web : Present and Future -Tushar Mhaskar Jayant Madhavan, Loredana Afanasiev, Lyublena Antova, Alon Halevy January 7,
GroceryList++ Brian, Peter, Staci, Darrin, Anna.
Neural Machine Translation
Scoring Sequence Alignments Calculating E
Approaches to Machine Translation
Language Technologies Institute Carnegie Mellon University
Chapter 6. Data Collection in a Wizard-of-Oz Experiment in Reinforcement Learning for Adaptive Dialogue Systems by: Rieser & Lemon. Course: Autonomous.
Joint Training for Pivot-based Neural Machine Translation
Suggestions for Class Projects
SECOND LANGUAGE LISTENING Comprehension: Process and Pedagogy
Approaches to Machine Translation
Statistical Machine Translation Papers from COLING 2004
LINGUA INGLESE 2A – a.a. 2018/2019 Computer-Aided Translation Technology LESSON 3 prof. ssa Laura Liucci –
Information Retrieval
Bug Localization with Combination of Deep Learning and Information Retrieval A. N. Lam et al. International Conference on Program Comprehension 2017.
Neural Machine Translation
Presentation transcript:

Data Driven Response Generation in Social Media Alan Ritter Colin Cherry Bill Dolan

Task: Response Generation Input: Arbitrary user utterance Output: Appropriate response Training Data: Millions of conversations from Twitter

Parallelism in Discourse (Hobbs 1985) STATUS: I am slowly making this soup and it smells gorgeous! RESPONSE: I’ll bet it looks delicious too!

Parallelism in Discourse (Hobbs 1985) STATUS: I am slowly making this soup and it smells gorgeous! RESPONSE: I’ll bet it looks delicious too!

Parallelism in Discourse (Hobbs 1985) STATUS: I am slowly making this soup and it smells gorgeous! RESPONSE: I’ll bet it looks delicious too!

Parallelism in Discourse (Hobbs 1985) STATUS: I am slowly making this soup and it smells gorgeous! RESPONSE: I’ll bet it looks delicious too!

Parallelism in Discourse (Hobbs 1985) STATUS: I am slowly making this soup and it smells gorgeous! RESPONSE: Can we “translate” the status into an appropriate response? I’ll bet it looks delicious too!

Why Should SMT work on conversations? Conversation and translation not the same Source and Target not Semantically Equivalent Can’t learn semantics behind conversations We Can learn some high-frequency patterns “I am” -> “you are” “airport” -> “safe flight” First step towards learning conversational models from data.

SMT: Advantages Leverage existing techniques Perform well Scalable Provides probabilistic model of responses Straightforward to integrate into applications

Data Driven Response Generation: Potential Applications Dialogue Generation (more natural responses)

Data Driven Response Generation: Potential Applications Dialogue Generation (more natural responses) Conversationally-aware predictive text entry Speech Interface to SMS/Twitter (Ju and Paek 2010) I’m feeling sick Status: Response: + = Hope you feel better Response:

Twitter Conversations Most of Twitter is broadcasting information: iPhone 4 on Verizon coming February 10th ..

Twitter Conversations Most of Twitter is broadcasting information: iPhone 4 on Verizon coming February 10th .. About 20% are replies I 'm going to the beach this weekend! Woo! And I'll be there until Tuesday. Life is good. Enjoy the beach! Hope you have great weather! thank you 

Data Crawled Twitter Public API 1.3 Million Conversations Easy to gather more data

No need for disentanglement Data Crawled Twitter Public API 1.3 Million Conversations Easy to gather more data No need for disentanglement (Elsner & Charniak 2008)

Approach: Statistical Machine Translation SMT Response Generation INPUT: Foreign Text User Utterance OUTPUT English Text Response TRAIN: Parallel Corpora Conversations

Approach: Statistical Machine Translation SMT Response Generation INPUT: Foreign Text User Utterance OUTPUT English Text Response TRAIN: Parallel Corpora Conversations

Phrase-Based Translation STATUS: who wants to come over for dinner tomorrow? RESPONSE:

Phrase-Based Translation STATUS: who wants to come over for dinner tomorrow? Yum ! I RESPONSE:

Phrase-Based Translation STATUS: who wants to come over for dinner tomorrow? Yum ! I want to RESPONSE:

Phrase-Based Translation STATUS: who wants to come over for dinner tomorrow? Yum ! I want to be there RESPONSE:

Phrase-Based Translation STATUS: who wants to come over for dinner tomorrow? Yum ! I want to be there tomorrow ! RESPONSE:

Phrase Based Decoding Log Linear Model Features Include: Language Model Phrase Translation Probabilities Additional feature functions…. Use Moses Decoder Beam Search

Challenges applying SMT to Conversation Wider range of possible targets Larger fraction of unaligned words/phrases Large phrase pairs which can’t be decomposed

Challenges applying SMT to Conversation Wider range of possible targets Larger fraction of unaligned words/phrases Large phrase pairs which can’t be decomposed Source and Target are not Semantically Equivelant

Challenge: Lexical Repetition Source/Target strings are in same language Strongest associations between identical pairs Without anything to discourage the use of lexically similar phrases, the system tends to “parrot back” input STATUS: I’m slowly making this soup ...... and it smells gorgeous! RESPONSE: I’m slowly making this soup ...... and you smell gorgeous!

Lexical Repitition: Solution Filter out phrase pairs where one is a substring of the other Novel feature which penalizes lexically similar phrase pairs Jaccard similarity between the set of words in the source and target

Word Alignment: Doesn’t really work… Typically used for Phrase Extraction GIZA++ Very poor alignments for Status/response pairs Alignments are very rarely one-to-one Large portions of source ignored Large phrase pairs which can’t be decomposed

Word Alignment Makes Sense Sometimes…

Sometimes Word Alignment is Very Difficult

Sometimes Word Alignment is Very Difficult Difficult Cases confuse IBM Word Alignment Models Poor Quality Alignments

Solution: Generate all phrase-pairs (With phrases up to length 4) Example: S: I am feeling sick R: Hope you feel better

Solution: Generate all phrase-pairs (With phrases up to length 4) Example: S: I am feeling sick R: Hope you feel better O(N*M) phrase pairs N = length of status M = length of response

Solution: Generate all phrase-pairs (With phrases up to length 4) Example: S: I am feeling sick R: Hope you feel better O(N*M) phrase pairs N = length of status M = length of response Source Target I Hope you feel … feeling sick feel better Hope you feel you feel better I am feeling

Pruning: Fisher Exact Test (Johson et. al. 2007) (Moore 2004) Details: Keep 5Million highest ranking phrase pairs Includes a subset of the (1,1,1) pairs Filter out pairs where one phrase is a substring

Example Phrase-Table Entries Source Target how are good wish me good luck sick feel better bed dreams interview how are you ? i 'm good to bed good night thanks for no problem r u i 'm my dad your dad airport have a safe can i you can

Baseline: Information Retrieval/ Nearest Neighbor (Swanson and Gordon 2008) (Isbell et. al. 2000) (Jafarpour and Burgess) Find the most similar response in training data 2 options to find response for status :

Mechanical Turk Evaluation Pairwise Comparison of Output (System A & B) For Each Experiment: Randomly select 200 status messages Generate response using systems A & B Ask Turkers which response is better Each HIT is submitted to 3 different workers

Results System A System B Fraction A Agreement (S) MT-Chat IR-Status 0.645 0.347 IR-Response 0.593 0.330 Human 0.145 0.433

Results Summary: System A System B Fraction A Agreement (S) MT-Chat IR-Status 0.645 0.347 IR-Response 0.593 0.330 Human 0.145 0.433 Summary: MT outperforms IR Direct comparison is better Looses to humans But, generates better response in 15% of cases

Cases where MT output was preferred Status MT-Chat Human I want to go to a bonfire right about now That does sound fun bonfire I wan na go! Did my tweet bring that upon you? hehe I’m the invisible man. Man, you could never be invisible I can still see your face surrounded in blue! Put ya number in my phone …and ima call you up Call me when u do that in your phone call  for what’s up? O well hw kum u dnt eva call or txt me ?!?! Lolz jpn hw r ya tho ?!?! Gray Maynard with hair? I did not approve this. Maynard with grey hair! I take it back the hair isn’t working for Maynard.

Demo www.cs.washington.edu/homes/aritter/mt_chat.html

Contributions Proposed SMT as an approach to Generating Responses Many Challenges in Adapting Phrase-Based SMT to Conversations Lexical Repetition Difficult Alignment Phrase-based translation performs better than IR Able to beat Human responses 15% of the time

Contributions Proposed SMT as an approach to Generating Responses Many Challenges in Adapting Phrase-Based SMT to Conversations Lexical Repetition Difficult Alignment Phrase-based translation performs better than IR Able to beat Human responses 15% of the time Thanks!

Phrase-Based Translation STATUS: who wants to get some lunch ? RESPONSE:

Phrase-Based Translation STATUS: who wants to get some lunch ? I wan na RESPONSE:

Phrase-Based Translation STATUS: who wants to get some lunch ? I wan na get me some RESPONSE:

Phrase-Based Translation STATUS: who wants to get some lunch ? I wan na get me some chicken RESPONSE: