Donna M. Gates Carnegie Mellon University

Slides:



Advertisements
Similar presentations
COGEX at the Second RTE Marta Tatu, Brandon Iles, John Slavick, Adrian Novischi, Dan Moldovan Language Computer Corporation April 10 th, 2006.
Advertisements

COGEX at the Second RTE Marta Tatu, Brandon Iles, John Slavick, Adrian Novischi, Dan Moldovan Language Computer Corporation April 10 th, 2006.
Recognizing Textual Entailment Challenge PASCAL Suleiman BaniHani.
For Friday No reading Homework –Chapter 23, exercises 1, 13, 14, 19 –Not as bad as it sounds –Do them IN ORDER – do not read ahead here.
Robust Textual Inference via Graph Matching Aria Haghighi Andrew Ng Christopher Manning.
Probabilistic Parsing: Enhancements Ling 571 Deep Processing Techniques for NLP January 26, 2011.
Predicting Cloze Task Quality for Vocabulary Training Adam Skory, Maxine Eskenazi Language Technologies Institute Carnegie Mellon University
PCFG Parsing, Evaluation, & Improvements Ling 571 Deep Processing Techniques for NLP January 24, 2011.
Introduction to Computational Linguistics Lecture 2.
Introduction to CL Session 1: 7/08/2011. What is computational linguistics? Processing natural language text by computers  for practical applications.
Information retrieval Finding relevant data using irrelevant keys Example: database of photographic images sorted by number, date. DBMS: Well structured.
Machine Learning in Natural Language Processing Noriko Tomuro November 16, 2006.
1 Basic Parsing with Context Free Grammars Chapter 13 September/October 2012 Lecture 6.
Finding Advertising Keywords on Web Pages Scott Wen-tau YihJoshua Goodman Microsoft Research Vitor R. Carvalho Carnegie Mellon University.
A Web-based Question Answering System Yu-shan & Wenxiu
AQUAINT Kickoff Meeting – December 2001 Integrating Robust Semantics, Event Detection, Information Fusion, and Summarization for Multimedia Question Answering.
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
Study of Automated Extraction of Security Policy from Natural-Language Software Documents * Nov. 21, 2013, Kaidi Ma, Man Sun Computer Information Science.
“How much context do you need?” An experiment about context size in Interactive Cross-language Question Answering B. Navarro, L. Moreno-Monteagudo, E.
For Friday Finish chapter 23 Homework: –Chapter 22, exercise 9.
The CoNLL-2013 Shared Task on Grammatical Error Correction Hwee Tou Ng, Yuanbin Wu, and Christian Hadiwinoto 1 Siew.
Using Text Mining and Natural Language Processing for Health Care Claims Processing Cihan ÜNAL
Scott Duvall, Brett South, Stéphane Meystre A Hands-on Introduction to Natural Language Processing in Healthcare Annotation as a Central Task for Development.
AQUAINT BBN’s AQUA Project Ana Licuanan, Jonathan May, Scott Miller, Ralph Weischedel, Jinxi Xu 3 December 2002.
21/11/2002 The Integration of Lexical Knowledge and External Resources for QA Hui YANG, Tat-Seng Chua Pris, School of Computing.
AQUAINT 18-Month Workshop 1 Light Semantic Processing for QA Language Technologies Institute, Carnegie Mellon B. Van Durme, Y. Huang, A. Kupsc and E. Nyberg.
14/12/2009ICON Dipankar Das and Sivaji Bandyopadhyay Department of Computer Science & Engineering Jadavpur University, Kolkata , India ICON.
Opinion Holders in Opinion Text from Online Newspapers Youngho Kim, Yuchul Jung and Sung-Hyon Myaeng Reporter: Chia-Ying Lee Advisor: Prof. Hsin-Hsi Chen.
GTRI.ppt-1 NLP Technology Applied to e-discovery Bill Underwood Principal Research Scientist “The Current Status and.
Natural Language Programming David Vadas The University of Sydney Supervisor: James Curran.
For Monday Read chapter 24, sections 1-3 Homework: –Chapter 23, exercise 8.
For Monday Read chapter 26 Last Homework –Chapter 23, exercise 7.
Automatic Question Answering  Introduction  Factoid Based Question Answering.
For Friday Finish chapter 23 Homework –Chapter 23, exercise 15.
Mining Dependency Relations for Query Expansion in Passage Retrieval Renxu Sun, Chai-Huat Ong, Tat-Seng Chua National University of Singapore SIGIR2006.
1 Evaluating High Accuracy Retrieval Techniques Chirag Shah,W. Bruce Croft Center for Intelligent Information Retrieval Department of Computer Science.
Answer Mining by Combining Extraction Techniques with Abductive Reasoning Sanda Harabagiu, Dan Moldovan, Christine Clark, Mitchell Bowden, Jown Williams.
11 Project, Part 3. Outline Basics of supervised learning using Naïve Bayes (using a simpler example) Features for the project 2.
FILTERED RANKING FOR BOOTSTRAPPING IN EVENT EXTRACTION Shasha Liao Ralph York University.
NLP. Introduction to NLP Last week, Min broke the window with a hammer. The window was broken with a hammer by Min last week With a hammer, Min broke.
1 Fine-grained and Coarse-grained Word Sense Disambiguation Jinying Chen, Hoa Trang Dang, Martha Palmer August 22, 2003.
Survey on Long Queries in Keyword Search : Phrase-based IR Sungchan Park
1 Question Answering and Logistics. 2 Class Logistics  Comments on proposals will be returned next week and may be available as early as Monday  Look.
For Monday Read chapter 26 Homework: –Chapter 23, exercises 8 and 9.
Overview of Statistical NLP IR Group Meeting March 7, 2006.
An Ontology-based Automatic Semantic Annotation Approach for Patent Document Retrieval in Product Innovation Design Feng Wang, Lanfen Lin, Zhou Yang College.
11 Thoughts on STS regarding Machine Reading Ralph Weischedel 12 March 2012.
Question Classification Ling573 NLP Systems and Applications April 25, 2013.
Automatic Ontology Extraction Miloš Husák RASLAN 2010.
Ensembling Diverse Approaches to Question Answering
Modeling Grammaticality
Basic Parsing with Context Free Grammars Chapter 13
Authorship Attribution Using Probabilistic Context-Free Grammars
Structured Browsing for Unstructured Text
Semantic Parsing for Question Answering
Natural Language Processing (NLP)
Two Discourse Driven Language Models for Semantics
Learning to Transform Natural to Formal Languages
Machine Learning in Natural Language Processing
Recognizing Partial Textual Entailment
Automatic Detection of Causal Relations for Question Answering
CS246: Information Retrieval
Natural Language Processing (NLP)
Modeling Grammaticality
The Winograd Schema Challenge Hector J. Levesque AAAI, 2011
Artificial Intelligence 2004 Speech & Natural Language Processing
Progress report on Semantic Role Labeling
Information Retrieval
Bug Localization with Combination of Deep Learning and Information Retrieval A. N. Lam et al. International Conference on Program Comprehension 2017.
Natural Language Processing (NLP)
Presentation transcript:

Donna M. Gates Carnegie Mellon University Generating Look-Back Reading Comprehension Questions from Expository Text Donna M. Gates Carnegie Mellon University September 25-26, 2008 Workshop on the Question Generation Shared Task and Evaluation Challenge

Fact Based Questions Reading Comprehension: Look-Back Strategy Question whose answer is Right-There-In-The-Text. (Raphael 1982) Automatic generation of questions: system needs to understand the text well enough to formulate questions. Q&A systems Questions are asked to find information. Need to find best matching answers or documents containing an answer to a query. (Leidner et al 2003) Solution: Annotate documents with syntactic and/or semantic information. September 25-26, 2008 Workshop on the Question Generation Shared Task and Evaluation Challenge

NLP Programs and Knowledge Sources Stanford NL Parser (Klein & Manning 2003) WordNet (Felbaum 1998) to get noun classifications BBN IdentiFinder (Bikel et al 1999) to get named entities Assert (Pradhan et al 2005) to produce PropBank (Palmer et al 2005) tags Stanford T-Surgeon (Levy & Galen 2006) + handwritten transformation rules to transform declarative sentence trees into question trees Code to combine the annotations and generate strings. September 25-26, 2008 Workshop on the Question Generation Shared Task and Evaluation Challenge

Example Text with Question September 25-26, 2008 Workshop on the Question Generation Shared Task and Evaluation Challenge

Evaluating Generated Questions Data Sets: Mitre’s CBC4Kids Q&A data 70+ texts for training 50+ texts for testing All automatically generated questions were evaluated for grammaticality/fluency and whether the answer that matched the question could be found easily. A single grader: More graders would be better. I could obtain intercoder agreement measures. Precision only Goal is to generate well-formed questions but goal is not to generate all possible questions from a single sentence. No gold-standard questions relevant for this specific task. September 25-26, 2008 Workshop on the Question Generation Shared Task and Evaluation Challenge

Evaluation Scoring Evaluation grades Acceptable = perfect + ok Perfect Mr. Yashin makes a salary of more than three million dollars a season. Who makes a salary of more than three million dollars a season? Ok A study was conducted by the aboriginals. Whom or what was conducted by the aboriginals? Bad Air-raid warning sirens sounded in the Kosovo capital of Pristina this morning. Who sounded in the Kosovo capital of Pristina this morning? WordNet ambiguities: siren (mythical type of person vs noise making device) Failed Memberships cost $180 a year for adults and $135 for students and seniors. Whom or what $180 did memberships cost a year for adults and $135 for students and seniors? Parsing problem Acceptable = perfect + ok September 25-26, 2008 Workshop on the Question Generation Shared Task and Evaluation Challenge

Results WH-Phrase Question Transformation Total Perfect OK Bad Failed Subj NP: Who conducted a study? 444 80% 6% 13% 2% Subj Gerund: What will be a new experience? (Conducting studies will be a new experience.) 0% D. Obj1: What did aboriginals conduct? (Aboriginals conducted a study.) 119 55% 8% 25% 12% D. Obj2: What were aboriginals conducting? 52 58% 10% D. Obj3: What will aboriginals conduct? 24 63% 17% Passive Ag: By whom were studies conducted? (Studies were conducted by aboriginals.) 11 91% 9% Temp: When did aboriginals conduct a study? PP/S: (On Friday aboriginals conducted a study.) 19 89% 5% .. NP/S: (Last Friday aboriginals conducted a study.) 1 100% PP/VP: (Aboriginals conducted a study on Friday.) 14 86% 7% NP/VP: (Aboriginals conducted a study last Friday.) 9 September 25-26, 2008 Workshop on the Question Generation Shared Task and Evaluation Challenge

Result Highlights Direct Object Wh Phrases Overall: 81% acceptable Subject Wh Phrases Aboriginals conducted a study last month. Who conducted a study last month? Largest number of examples (444): 86% acceptable. Direct Object Wh Phrases What did aboriginals conduct last month? What will aboriginals conduct? What were aboriginals conducting? Combined, lowest acceptable scores: 66% acceptable Wh NP Temporal Expressions: last month When did aboriginals conduct a study? 100% perfect (10 parsed and annotated correctly) September 25-26, 2008 Workshop on the Question Generation Shared Task and Evaluation Challenge

Issues to be Resolved Need to expand semantic annotation and transformations to include locations. Improve use of WordNet by filtering low frequency senses: dish (satellite dish vs attractive person) - WHO vs WHAT. Incorporate other syntactic and semantic annotators. Define a gold standard/target set of questions. September 25-26, 2008 Workshop on the Question Generation Shared Task and Evaluation Challenge