Automatic Question Answering  Introduction  Factoid Based Question Answering.

Slides:



Advertisements
Similar presentations
Improved TF-IDF Ranker
Advertisements

Recognizing Textual Entailment Challenge PASCAL Suleiman BaniHani.
SEARCHING QUESTION AND ANSWER ARCHIVES Dr. Jiwoon Jeon Presented by CHARANYA VENKATESH KUMAR.
Multimedia Answer Generation for Community Question Answering.
TEMPLATE DESIGN © Identifying Noun Product Features that Imply Opinions Lei Zhang Bing Liu Department of Computer Science,
Query Dependent Pseudo-Relevance Feedback based on Wikipedia SIGIR ‘09 Advisor: Dr. Koh Jia-Ling Speaker: Lin, Yi-Jhen Date: 2010/01/24 1.
Vikas BhardwajColumbia University NLP for the Web – Spring 2010 Improving QA Accuracy by Question Inversion Prager et al. IBM T.J. Watson Res. Ctr. 02/18/2010.
Aki Hecht Seminar in Databases (236826) January 2009
The Informative Role of WordNet in Open-Domain Question Answering Marius Paşca and Sanda M. Harabagiu (NAACL 2001) Presented by Shauna Eggers CS 620 February.
Basi di dati distribuite Prof. M.T. PAZIENZA a.a
Information Extraction and Ontology Learning Guided by Web Directory Authors:Martin Kavalec Vojtěch Svátek Presenter: Mark Vickers.
Writing Good Software Engineering Research Papers A Paper by Mary Shaw In Proceedings of the 25th International Conference on Software Engineering (ICSE),
Article by: Feiyu Xu, Daniela Kurz, Jakub Piskorski, Sven Schmeier Article Summary by Mark Vickers.
Employing Two Question Answering Systems in TREC 2005 Harabagiu, Moldovan, et al 2005 Language Computer Corporation.
Enhance legal retrieval applications with an automatically induced knowledge base Ka Kan Lo.
Chapter 5: Information Retrieval and Web Search
Learning Table Extraction from Examples Ashwin Tengli, Yiming Yang and Nian Li Ma School of Computer Science Carnegie Mellon University Coling 04.
CSC 9010 Spring Paula Matuszek A Brief Overview of Watson.
Hang Cui et al. NUS at TREC-13 QA Main Task 1/20 National University of Singapore at the TREC- 13 Question Answering Main Task Hang Cui Keya Li Renxu Sun.
A Technical Seminar on Question Answering SHRI RAMDEOBABA COLLEGE OF ENGINEERING & MANAGEMENT Presented By: Rohini Kamdi Guided By: Dr. A.J.Agrawal.
Going Beyond Simple Question Answering Bahareh Sarrafzadeh CS 886 – Spring 2015.
AnswerBus Question Answering System Zhiping Zheng School of Information, University of Michigan HLT 2002.
Question Answering.  Goal  Automatically answer questions submitted by humans in a natural language form  Approaches  Rely on techniques from diverse.
Question Answering From Zero to Hero Elena Eneva 11 Oct 2001 Advanced IR Seminar.
Incident Threading for News Passages (CIKM 09) Speaker: Yi-lin,Hsu Advisor: Dr. Koh, Jia-ling. Date:2010/06/14.
A Probabilistic Graphical Model for Joint Answer Ranking in Question Answering Jeongwoo Ko, Luo Si, Eric Nyberg (SIGIR ’ 07) Speaker: Cho, Chin Wei Advisor:
Carnegie Mellon School of Computer Science Copyright © 2001, Carnegie Mellon. All Rights Reserved. JAVELIN Project Briefing 1 AQUAINT Phase I Kickoff December.
Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.
11 A Hybrid Phish Detection Approach by Identity Discovery and Keywords Retrieval Reporter: 林佳宜 /10/17.
QUALIFIER in TREC-12 QA Main Task Hui Yang, Hang Cui, Min-Yen Kan, Mstislav Maslennikov, Long Qiu, Tat-Seng Chua School of Computing National University.
A Language Independent Method for Question Classification COLING 2004.
21/11/2002 The Integration of Lexical Knowledge and External Resources for QA Hui YANG, Tat-Seng Chua Pris, School of Computing.
Giorgos Giannopoulos (IMIS/”Athena” R.C and NTU Athens, Greece) Theodore Dalamagas (IMIS/”Athena” R.C., Greece) Timos Sellis (IMIS/”Athena” R.C and NTU.
INTERESTING NUGGETS AND THEIR IMPACT ON DEFINITIONAL QUESTION ANSWERING Kian-Wei Kor, Tat-Seng Chua Department of Computer Science School of Computing.
Chapter 6: Information Retrieval and Web Search
AQUAINT 18-Month Workshop 1 Light Semantic Processing for QA Language Technologies Institute, Carnegie Mellon B. Van Durme, Y. Huang, A. Kupsc and E. Nyberg.
PIQUANT at AQUAINT Kick-Off Dec PIQUANT Practical Intelligent QUestion ANswering Technology A Question Answering system integrating Information.
EasyQuerier: A Keyword Interface in Web Database Integration System Xian Li 1, Weiyi Meng 2, Xiaofeng Meng 1 1 WAMDM Lab, RUC & 2 SUNY Binghamton.
Enhancing Cluster Labeling Using Wikipedia David Carmel, Haggai Roitman, Naama Zwerdling IBM Research Lab (SIGIR’09) Date: 11/09/2009 Speaker: Cho, Chin.
A Novel Pattern Learning Method for Open Domain Question Answering IJCNLP 2004 Yongping Du, Xuanjing Huang, Xin Li, Lide Wu.
LOGO 1 Corroborate and Learn Facts from the Web Advisor : Dr. Koh Jia-Ling Speaker : Tu Yi-Lang Date : Shubin Zhao, Jonathan Betz (KDD '07 )
Authors: Marius Pasca and Benjamin Van Durme Presented by Bonan Min Weakly-Supervised Acquisition of Open- Domain Classes and Class Attributes from Web.
©2003 Paula Matuszek CSC 9010: Text Mining Applications Dr. Paula Matuszek (610)
Distribution of information in biomedical abstracts and full- text publications M. J. Schuemie et al. Dept. of Medical Informatics, Erasmus University.
Using a Named Entity Tagger to Generalise Surface Matching Text Patterns for Question Answering Mark A. Greenwood and Robert Gaizauskas Natural Language.
Conceptual structures in modern information retrieval Claudio Carpineto Fondazione Ugo Bordoni
August 17, 2005Question Answering Passage Retrieval Using Dependency Parsing 1/28 Question Answering Passage Retrieval Using Dependency Parsing Hang Cui.
A Classification-based Approach to Question Answering in Discussion Boards Liangjie Hong, Brian D. Davison Lehigh University (SIGIR ’ 09) Speaker: Cho,
AQUAINT IBM PIQUANT ARDACYCORP Subcontractor: IBM Question Answering Update piQuAnt ARDA/AQUAINT December 2002 Workshop This work was supported in part.
Date: 2012/08/21 Source: Zhong Zeng, Zhifeng Bao, Tok Wang Ling, Mong Li Lee (KEYS’12) Speaker: Er-Gang Liu Advisor: Dr. Jia-ling Koh 1.
Mining Dependency Relations for Query Expansion in Passage Retrieval Renxu Sun, Chai-Huat Ong, Tat-Seng Chua National University of Singapore SIGIR2006.
Introduction to Data Mining by Yen-Hsien Lee Department of Information Management College of Management National Sun Yat-Sen University March 4, 2003.
Answer Mining by Combining Extraction Techniques with Abductive Reasoning Sanda Harabagiu, Dan Moldovan, Christine Clark, Mitchell Bowden, Jown Williams.
Extracting and Ranking Product Features in Opinion Documents Lei Zhang #, Bing Liu #, Suk Hwan Lim *, Eamonn O’Brien-Strain * # University of Illinois.
AQUAINT AQUAINT Evaluation Overview Ellen M. Voorhees.
Survey Jaehui Park Copyright  2008 by CEBT Introduction  Members Jung-Yeon Yang, Jaehui Park, Sungchan Park, Jongheum Yeon  We are interested.
Keyword Translation Accuracy and Cross-Lingual Question Answering in Chinese and Japanese Teruko Mitamura Mengqiu Wang Hideki Shima Frank Lin In CMU EACL.
SEMANTIC VERIFICATION IN AN ONLINE FACT SEEKING ENVIRONMENT DMITRI ROUSSINOV, OZGUR TURETKEN Speaker: Li, HueiJyun Advisor: Koh, JiaLing Date: 2008/5/1.
Survey on Long Queries in Keyword Search : Phrase-based IR Sungchan Park
1 Question Answering and Logistics. 2 Class Logistics  Comments on proposals will be returned next week and may be available as early as Monday  Look.
1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis.
Instance Discovery and Schema Matching With Applications to Biological Deep Web Data Integration Tantan Liu, Fan Wang, Gagan Agrawal {liut, wangfa,
Question Answering Passage Retrieval Using Dependency Relations (SIGIR 2005) (National University of Singapore) Hang Cui, Renxu Sun, Keya Li, Min-Yen Kan,
University Of Seoul Ubiquitous Sensor Network Lab Query Dependent Pseudo-Relevance Feedback based on Wikipedia 전자전기컴퓨터공학 부 USN 연구실 G
Do-Gil Lee1*, Ilhwan Kim1 and Seok Kee Lee2
Web IR: Recent Trends; Future of Web Search
Traditional Question Answering System: an Overview
Presentation 王睿.
CSc4730/6730 Scientific Visualization
Tantan Liu, Fan Wang, Gagan Agrawal The Ohio State University
Presentation transcript:

Automatic Question Answering  Introduction  Factoid Based Question Answering

Uses of Automatic Question Answering FAQ (Frequently Asked Questions) Help Desks / Customer Service Phone Centers Accessing Complex set of Technical Maintenance Manuals Integrating QA in Knowledge Management and Portals Wide variety of Other E-Business Applications

Introduction

Evaluating Extracted Answers

A Brief Overview of QA systems

General architecture question Question Classification/ Preprocessing Information Retrieval Answer Extraction Answeranswer e.g. What is Calvados? /Q is /A where:/Q=“( Calvados)” Query=“Calvados is” Text retrieva l=“…Calvados is often used in cooking…Calvado s is a dry apple brandy made in… /A is : a dry apple brandy Answer: /Q is /A: “Calvados ” is ”a dry apple brandy”

Approaches: Structured Knowledge-Base Approach Deepest QA but Limited to Given Subject Domain Create comprehensive Knowledge Base(s) or other Structured Data Base(s) At the 10K Axiom Level -- Capable of Answering factual questions within domain At the 100K Axiom Level -- Answer cause & effect/capability Questions At the 1000K Axiom Level -- Answer Novel Questions; ID alternatives Ref: Dr. John D. Prange; AAAI Symposium 2002

Different Approaches Question: What is the largest city in England? Text Match – Find text that says “London is the largest city in England” (or paraphrase). Confidence is confidence of NL parser * confidence of source. Find multiple instances and confidence of source -> 1. “Superlative” Search – Find a table of English cities and their populations, and sort. – Find a list of the 10 largest cities in the world, and see which are in England. Uses logic: if L > all objects in set R then L > all objects in set E < R. – Find the population of as many individual English cities as possible, and choose the largest. Heuristics – London is the capital of England. (Not guaranteed to imply it is the largest city, but this is very frequently the case.) Complex Inference – E.g. “Birmingham is England’s second-largest city”; “Paris is larger than Birmingham”; “London is larger than Paris”; “London is in England”. Reference: IBM TJ Watson AQUAINT Briefing

Problem Statements and supporting approaches

Named-Entity (NE) based approach The idea is that factoid questions fall into several distinctive types, such as “location”, “date”, “person”, etc. Assuming that we can recognize the question type correctly, then the potential answer candidates can be limited down to a few NE types that correspond to the question type. – Intuitively, if the question is asking for a date, then an answer string that is identified to be a location type named-entity is not likely to be the correct answer. However, it is important to bear in mind that neither question type classification nor NE recognition are perfect in the real world. Such systems can be harmful when classification and recognition errors occur. The aforementioned two types of information – sentence surface text patterns and answer candidate NE type – both come from the answer sentence side. The only information we have extracted from the question side is the question type, which is used for selecting patterns and NE types for matching. – The structural information in the question sentence, which is not available in inputs in many other tasks (e.g. ad-hoc document retrieval), has not yet been fully-utilized. Recognizing this unique input source information, there have been many recent work on finding ways to better utilize structural information, and we will review them next.

Models for combining multiple sources of information in answer scoring Nearly all QA systems use evidence from multiple sources to decide what the best answer is for a given question. Lee et al.(2005) first decided the NE type of answer candidates based on the question type, and then ranked answer candidates based on four types of evidence: 1.The count of NE types that occurred in the question occurring in the answer sentence, normalized by the number of NE types occurred in the question. 2.The count of constraints other than NE type constraint that occurred in the question occurring in the answer sentence, normalized by the number of constraints occurred in the question. 3.A pre-set score when the answer candidate string contains the identified question focus string. 4.A pre-set score the answer candidate string appear in the answer sentence next to a term that contains the question focus string. Each listed condition will give a normalized score if satisfied, and the final scoring function is the sum of these scores.

Ad-Hoc Feature Weighting Another example of ad-hoc feature weighting is the TREC-9 system described in (Ittycheriah et al. 2001). In the answer selection module, the following distance metrics are computed for each 3 sentences window: 1.Matching words: the sum of TFIDF scores of the question words appear in the 3 sentence window. 2.Thesaurus match: the sum of TFIDF scores of the question words whose thesaurus matching by WordNet appear in the 3 sentence window. 3.Mis-matching words: the sum of TFIDF scores of the question words that do not appear in the 3 sentence window. 4.Dispersion: the number of words in the answer sentence that occur between matched question words. 5.Cluster words: the number of words that occur next to each other in both the answer sentence window and the question Having computed these distance metrics, they used an ad-hoc scoring function which was not detailed in the paper to combine these metrics.

Reference Mengqiu Wang; A Survey of Answer Extraction Techniques in Factoid Question Answering; Association for Computational Linguistics; M. W. Bilotti. Linguistic and Semantic Passage Retrieval Strategies for Question Answering. PhD thesis, Carnegie Mellon University, M. W. Bilotti, P. Ogilvie, J. Callan, and E. Nyberg. Structured retrieval for question answering. In Proc. SIGIR, H. Cui, R. Sun, K. Li, M.-Y. Kan, and T.-S. Chua. Question answering passage retrieval using dependency relations. In Proc. SIGIR, X. Geng, T. Liu, T. Qin, and H. Li. Feature selection for ranking. In Proc. SIGIR, M. Kaisser and J. B. Lowe. Creating a research collection of question answer sentence pairs with amazon’s mechanical turk. In Proc. LREC, D. Shen and M. Lapata. Using semantic roles to improve question answering. In Proc. EMNLP, 2007.