Automatic Answer Validation in Open-Domain Question Answering Hristo Tanev TCC,ITC - IRST.

Slides:



Advertisements
Similar presentations
Improved TF-IDF Ranker
Advertisements

Semantic Access to Data from the Web Raquel Trillo *, Laura Po +, Sergio Ilarri *, Sonia Bergamaschi + and E. Mena * 1st International Workshop on Interoperability.
SEARCHING QUESTION AND ANSWER ARCHIVES Dr. Jiwoon Jeon Presented by CHARANYA VENKATESH KUMAR.
Natural Language Processing WEB SEARCH ENGINES August, 2002.
Group Members: Satadru Biswas ( ) Tanmay Khirwadkar ( ) Arun Karthikeyan Karra (05d05020) CS Course Seminar Group-2 Question Answering.
Query Operations: Automatic Local Analysis. Introduction Difficulty of formulating user queries –Insufficient knowledge of the collection –Insufficient.
Basic IR: Queries Query is statement of user’s information need. Index is designed to map queries to likely to be relevant documents. Query type, content,
Chapter 5: Query Operations Baeza-Yates, 1999 Modern Information Retrieval.
Anatomy of a Large-Scale Hypertextual Web Search Engine (e.g. Google)
6/16/20151 Recent Results in Automatic Web Resource Discovery Soumen Chakrabartiv Presentation by Cui Tao.
The Informative Role of WordNet in Open-Domain Question Answering Marius Paşca and Sanda M. Harabagiu (NAACL 2001) Presented by Shauna Eggers CS 620 February.
1 QA in Discussion Boards  Companies (e.g., Dell, IBM) use discussion boards as ways for customers to get answers to their questions  90% of 40 analyzed.
Computer comunication B Information retrieval. Information retrieval: introduction 1 This topic addresses the question on how it is possible to find relevant.
1 Information Retrieval and Extraction 資訊檢索與擷取 Chia-Hui Chang, Assistant Professor Dept. of Computer Science & Information Engineering National Central.
Information Retrieval and Extraction 資訊檢索與擷取 Chia-Hui Chang National Central University
Information retrieval Finding relevant data using irrelevant keys Example: database of photographic images sorted by number, date. DBMS: Well structured.
Employing Two Question Answering Systems in TREC 2005 Harabagiu, Moldovan, et al 2005 Language Computer Corporation.
Enhance legal retrieval applications with an automatically induced knowledge base Ka Kan Lo.
Use of Patterns for Detection of Answer Strings Soubbotin and Soubbotin.
Overview of Search Engines
 Search engines are programs that search documents for specified keywords and returns a list of the documents where the keywords were found.  A search.
Web Searching. Web Search Engine A web search engine is designed to search for information on the World Wide Web and FTP servers The search results are.
Information Retrieval in Practice
Probabilistic Model for Definitional Question Answering Kyoung-Soo Han, Young-In Song, and Hae-Chang Rim Korea University SIGIR 2006.
A Technical Seminar on Question Answering SHRI RAMDEOBABA COLLEGE OF ENGINEERING & MANAGEMENT Presented By: Rohini Kamdi Guided By: Dr. A.J.Agrawal.
Information Need Question Understanding Selecting Sources Information Retrieval and Extraction Answer Determina tion Answer Presentation This work is supported.
“How much context do you need?” An experiment about context size in Interactive Cross-language Question Answering B. Navarro, L. Moreno-Monteagudo, E.
AnswerBus Question Answering System Zhiping Zheng School of Information, University of Michigan HLT 2002.
Question Answering.  Goal  Automatically answer questions submitted by humans in a natural language form  Approaches  Rely on techniques from diverse.
Michael Cafarella Alon HalevyNodira Khoussainova University of Washington Google, incUniversity of Washington Data Integration for Relational Web.
When Experts Agree: Using Non-Affiliated Experts To Rank Popular Topics Meital Aizen.
Flexible Text Mining using Interactive Information Extraction David Milward
A Probabilistic Graphical Model for Joint Answer Ranking in Question Answering Jeongwoo Ko, Luo Si, Eric Nyberg (SIGIR ’ 07) Speaker: Cho, Chin Wei Advisor:
Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.
Natural Language Based Reformulation Resource and Web Exploitation for Question Answering Ulf Hermjakob, Abdessamad Echihabi, Daniel Marcu University of.
A Language Independent Method for Question Classification COLING 2004.
21/11/2002 The Integration of Lexical Knowledge and External Resources for QA Hui YANG, Tat-Seng Chua Pris, School of Computing.
TOPIC CENTRIC QUERY ROUTING Research Methods (CS689) 11/21/00 By Anupam Khanal.
The Internet 8th Edition Tutorial 4 Searching the Web.
Terminology and documentation*  Object of the study of terminology:  analysis and description of the units representing specialized knowledge in specialized.
Introduction to Digital Libraries hussein suleman uct cs honours 2003.
Comparing and Ranking Documents Once our search engine has retrieved a set of documents, we may want to Rank them by relevance –Which are the best fit.
Search Engine Architecture
Collocations and Information Management Applications Gregor Erbach Saarland University Saarbrücken.
Searching the web Enormous amount of information –In 1994, 100 thousand pages indexed –In 1997, 100 million pages indexed –In June, 2000, 500 million pages.
4 1 SEARCHING THE WEB Using Search Engines and Directories Effectively New Perspectives on THE INTERNET.
AQUAINT Kickoff Meeting Advanced Techniques for Answer Extraction and Formulation Language Computer Corporation Dallas, Texas.
Personalization with user’s local data Personalizing Search via Automated Analysis of Interests and Activities 1 Sungjick Lee Department of Electrical.
Evaluation of (Search) Results How do we know if our results are any good? Evaluating a search engine  Benchmarks  Precision and recall Results summaries:
Search Tools and Search Engines Searching for Information and common found internet file types.
Department of Software and Computing Systems Research Group of Language Processing and Information Systems The DLSIUAES Team’s Participation in the TAC.
Automatic Question Answering  Introduction  Factoid Based Question Answering.
Comparing Document Segmentation for Passage Retrieval in Question Answering Jorg Tiedemann University of Groningen presented by: Moy’awiah Al-Shannaq
Mining Dependency Relations for Query Expansion in Passage Retrieval Renxu Sun, Chai-Huat Ong, Tat-Seng Chua National University of Singapore SIGIR2006.
Answer Mining by Combining Extraction Techniques with Abductive Reasoning Sanda Harabagiu, Dan Moldovan, Christine Clark, Mitchell Bowden, Jown Williams.
Acquisition of Categorized Named Entities for Web Search Marius Pasca Google Inc. from Conference on Information and Knowledge Management (CIKM) ’04.
Relevance Models and Answer Granularity for Question Answering W. Bruce Croft and James Allan CIIR University of Massachusetts, Amherst.
SEMANTIC VERIFICATION IN AN ONLINE FACT SEEKING ENVIRONMENT DMITRI ROUSSINOV, OZGUR TURETKEN Speaker: Li, HueiJyun Advisor: Koh, JiaLing Date: 2008/5/1.
1 Question Answering and Logistics. 2 Class Logistics  Comments on proposals will be returned next week and may be available as early as Monday  Look.
Question Answering Passage Retrieval Using Dependency Relations (SIGIR 2005) (National University of Singapore) Hang Cui, Renxu Sun, Keya Li, Min-Yen Kan,
General Architecture of Retrieval Systems 1Adrienn Skrop.
Designing Cross-Language Information Retrieval System using various Techniques of Query Expansion and Indexing for Improved Performance  Hello everyone,
Search Engines and Search techniques
Types of Search Questions
Multimedia Information Retrieval
Search Techniques and Advanced tools for Researchers
Introduction to Information Retrieval
The ultimate in data organization
Information Retrieval and Web Design
Presentation transcript:

Automatic Answer Validation in Open-Domain Question Answering Hristo Tanev TCC,ITC - IRST

Open Domain Question Answering Automatic extracting of the answer of a natural language question Related fields: –Information Extraction –Information Retrieval Deeper text analysis Which is the capital of Italy? ROME

How it works

Question processing Which is the capital of Italy ? Question type: Which-LOCATION Keywords:capital Italy

IR engine Document collection Selected documents /paragraph s Question type: Which-LOCATION Keywords:capital Italy

Answer extraction Selected documents/ paragraphs Candidate answers Paris, Milan, Rome, Texas Question type: Which-LOCATION Keywords:capital Italy

Answer evaluation and validation Candidate answers Knowledge bases, Abduction Paris - 1, Milan - 1, Rome - 2

ROME!

The complexity of the QA task The variety of question classes The infinite number of answer formulations Anaphora, ellipsis, synonymy Sometimes syntactic and semantic analysis are necessary, also world knowledge

Answer inference The problem: How to infer if a candidate answer is relevant with respect to the question? Filtering out the irrelevant answer candidates Score the candidate answers according to their relevance

Contemporary approaches for answer inference Deducing the question logical form (QLF) from the text logical form Sanda M. Harabagiu and Marius Pasca and Steven Maiorano “Experiments with Open-Domain Textual Question Answering”, COLING 2000, Example: Q: Why did David Koresh ask FBI for a word processor? A: Mr. Koresh sent a request for word processor to FBI to enable him to write his revelations QLF: ask(Koresh, FBI, word processor,reason=?) ALF: sent request(Koresh, FBI, word processor,reason: to write his revelations) Heuristic: send request => ask, ALF => QLF

Contemporary approaches for answer inference (continued) Abduction, using pragmatic axioms, and semantic representation Sanda Harabagiu, Steven Maiorano “Finding Answers in Large Collections of Texts: Paragraph Indexing + Abductive Inference ” –action1 (e1, Person1) & action1 (e2, Person2) & related_events(e1, e2) => related(Person1, Person2) –Q:Who was Lincoln’s Secretary of State –A:Booth schemed to kill Lincoln, while his compatriots would murder Vice President Andrew Johnson and Secretary of State William Seward. –kill(e1, Lincoln) & murder(e2, Secretary of State William Seward) & related(e1, e2) =>related(Lincoln, Secretary of State William Seward)

Contemporary approaches for answer inference (continued) Lexico – syntactic patterns Q: What forms of international crime exist? A:…international forms of crime, including terrorism, blackmail and drug-related problems. These kinds of patterns are appropriate for certain type of questions, asking for taxonomic information

Contemporary approaches for answer inference - disadvantages A very large open domain knowledge base is requisite The creation of knowledge bases is very expensive in time and resources The present world knowledge bases (such as WordNet or ThoughtTreasure) are far away from being comprehensive The question and its answer can be very different lexically, this poses the necessity from deep semantic analysis to infer the relation between the discourse entities

Data Driven Answer Inference

The simple approach – ask the oracle Rome is the capital of Italy The database should be large enough to encode a great part of the human knowledge It should provide the necessary redundancy to contain different reformulations for the facts It should be changed dynamically to reflect recent state of the human knowledge about the world It should be easily accessible

World Wide Web as a source of knowledge Comprehensive Open domain nature Constantly updated and expanded Search indices and engines Implicit knowledge My journey in Italy began in the capital Rome… Disadvantages: Knowledge is in unstructured text form Access to the search engines may be slow

Web as a gigantic corpus Parameters: – hosts –AltaVista indexes over Web pages –Google Web pages –86% English language pages, 5.8% German, 2.36 French, 1.6% Italian Accessibility Different public accessible search engines: AltaVista, Fast, Google, Excite, Lycos, Yahoo!, Northern Light

Validation Statements Question Candidate Answer Who is Galileo? astronomer Galileo is an astronomer Which is the capital of Italy? Rome Rome is the capital of Italy Why the moon turns orange? because it enters the Earth shadow The moon turns orange because it enters the Earth shadow

Validation Statements (continued) The core of the data-driven answer validation is searching on-line texts, similar to the validation statement for a question- answer pair

The Answer Validation Algorithm Question + Answer = Validation Pattern –Q: How far is it from Denver to Aspen –A: 200 miles –QAP : [Denver … Aspen … 200 miles] Submit validation pattern to search engine Infer the power of the relation between Question & Answer on the basis of the search engine result

An Example QA pair: Who is Galileo? astronomer Submit to AltaVista the query “Galileo” –AltaVista returns 2000 hits Submit to AltaVista the query “astronomer” –AltaVista returns hits Submit to AltaVista the query Galileo NEAR astronomer –AltaVista returns 1000 hits PMI(Galileo, astronomer) = 14 > threshold

Validation Patterns The validation pattern is the base of the query which is submitted to the search engine to check if the question and the answer tend to appear together

Word Level Validation Patterns Qk1, Qk2,…. The question keywords A The Answer The query to the search engine is formed by linking the question keywords and answer with operators like AND or NEAR –Qk1 NEAR Qk2 NEAR …NEAR A –Qk1 AND Qk2 AND …AND A –(Qk1 AND Qk2 …) NEAR A This way co-occurrence between the question and answer keywords is searched in Intenet

Phrase Level Validation Patterns Validation pattern is composed by syntactic phrases instead of separate keywords Example: –Q: What city had a world fair in 1900? –A: Paris –Query: ( city NEAR “world fair” NEAR “in 1900” ) NEAR Paris Pages found by these type of patterns are more likely to contain texts confirming the answer corectness Disadvantage: less probable, often obtain 0 hits even for the right answer

Phrase Level Validation Patterns (continued) The phrases may be extracted by parser from the question More probable and coherent phrases should be preferred over the rare and non coherent phrases The phrase frequency may be measured using Web as a corpus

Sentence Level Patterns If the question and the answer are short, the whole validation statement can be submitted to the search engine “When did Hawaii become a state?” – 1959 “Hawaii became a state in 1959” Linguistic transformations are necessary to transform the QA pair in a validation statement

Morphological Variations and Symonymy in Patterns The question and answer keywords may occur in different morphological forms Synonyms can also appear instead of the original keywords Most search engines (Google, AltaVista, Yahoo) allow the use of keyword variants by OR operator Q: What date did John Lenon die? Question pattern: John NEAR Lenon NEAR (die OR died)

Types of data driven answer inference Pure quantitative approach: only the number of hits, returned by the search engine are considered. Statistical techniques form the core of this class of approaches Qualitative approaches: the document content is processed

Statistical answer validation By search engine queries are obtained the frequencies of the question pattern, the answer and the question- answer validation pattern Example Question: How far is it from Denver to Aspen? Question Pattern: far NEAR Denver NEAR Aspen Answer: 200 miles QAP: far NEAR Denver NEAR Aspen NEAR 200 miles Search engine: –Frequency(Question Pattern) –Frequency(Answer) –Frequency(QAP)

Statistical answer validation Using the frequencies and the number approximating the pages indexed by the search engine are calculated the following probabilities for occurrence in Web: P(Question Pattern), P(Answer), P(Question-Answer co-occurrence)

Statistical answer validation Thus calculated probabilities are combined in formulae, which are derived from classical co- occurrence formulae. The difference from the classical co-occurrence task is that we search how the appearance of the question pattern implies the appearance of the answer. Thus non symmetrical formulae are necessary. These formulae return a value, which is an indication for the answer corectness with respect to the question.

Statistical answer validation Answer validation formulae

Qualitative Approach The qualitative answer validation considers the content of the obtained documents as a result of the validation pattern submition to the search engine The distance between the question and answer keywords is considered

Qualitative Approach The use of document snippets can speed up this approach Certain search engines, like Yahoo! and Google return text snippets from the documents, where the keywords appear

Qualitative Approach. Extraction of data from the snippets. Q: Who is the first man to fly across the Pacific Ocean? A: Pangborn Query, submitted to Google: first AND man AND fly OR flew AND Pacific AND Ocean AND Pangborn Text snippets returned: “Pangborn became the first pilot to cross Pacific” “Pangborn with co-pilot Hew Herndon flew across Pacific”

Qualitative Approach. Extraction of data from the snippets (continued). Obtained co-occurrence relations: (Pangborn, first, Pacific) (Pangborn, fly, Pacific) Numerical values, obtained from the relations: Proportion of question keywords, related to answer (0.6 in the example, 3 question keywords (first, Pacific, fly) related to answer from total of 5 question keywords) Number of different relations and their length

Qualitative Approach. Calculating answer relevance Only the different co-occurrence relations are considered, co-occurrences, which are included in others are excluded PQK percent of question keywords, related to the answer r relations, obtained for the answer from the snippets length(r)the number of words in the co-occurrence relation r

Qualitative Approach. Calculating answer relevance (continued) Keyword density in the co-occurrence relations may also be considered The formula may be the sum of the keyword densities for all the relations

Combining approaches The qualitative approach can be used to extract co- occurrences Statistical techniques can be used to evaluate these co- ocurrences

Experiments and results

Experiment The statistical approach was tested The TREC10 question-answer list has been used, provided by NIST –for total of 492 questions maximum three right and three wrong answers are taken Two experiments were carried out Performance of the system on the full set of questions Named entities questions A baseline model was introduced

Experiment (continued) For every 50 byte answer the algorithm extracts only the entities that correspond to the question type The pairs question – answer were evaluated using AltaVista Phrase level patterns and two types of word-level patterns has been used

Experiment. The Patterns. Three types of patterns: –Phrase,Word level with NEAR,Word level with AND Example: –Q: “What city had a world fair in 1900?” –A: Paris –Phrase pattern: –(city NEAR “world fair” NEAR 1900) NEAR Paris –Word level with NEAR: –(city NEAR world NEAR fair NEAR 1900) NEAR Paris –World level with AND: –(city AND world AND fair AND 1900) NEAR Paris

Results Test SetSuccess Rate 3000 question-answer pairs from TREC10 81% 1500 question-answer pairs for named entity questions from TREC10 86% Baseline model52%

Future Directions

Much more to do… Improvement of the statistical formulae Research on the search engine use Combining the qualitative and statistical approach Creation of reliable validation patterns Introducing new techniques for answer validation Integration in QA system