Information Retrieval and Question-Answering

Slides:



Advertisements
Similar presentations
Chapter 5: Introduction to Information Retrieval
Advertisements

Introduction to Information Retrieval
Crawling, Ranking and Indexing. Organizing the Web The Web is big. Really big. –Over 3 billion pages, just in the indexable Web The Web is dynamic Problems:
Group Members: Satadru Biswas ( ) Tanmay Khirwadkar ( ) Arun Karthikeyan Karra (05d05020) CS Course Seminar Group-2 Question Answering.
Query Dependent Pseudo-Relevance Feedback based on Wikipedia SIGIR ‘09 Advisor: Dr. Koh Jia-Ling Speaker: Lin, Yi-Jhen Date: 2010/01/24 1.
Introduction to Information Retrieval (Manning, Raghavan, Schutze) Chapter 6 Scoring term weighting and the vector space model.
Information Retrieval in Practice
Information Retrieval Ling573 NLP Systems and Applications April 26, 2011.
Search and Retrieval: More on Term Weighting and Document Ranking Prof. Marti Hearst SIMS 202, Lecture 22.
Basic IR: Queries Query is statement of user’s information need. Index is designed to map queries to likely to be relevant documents. Query type, content,
The Informative Role of WordNet in Open-Domain Question Answering Marius Paşca and Sanda M. Harabagiu (NAACL 2001) Presented by Shauna Eggers CS 620 February.
Ch 4: Information Retrieval and Text Mining
Web-based Factoid Question Answering (including a sketch of Information Retrieval) Slides adapted from Dan Jurafsky, Jim Martin and Ed Hovy.
1 Question-Answering via the Web: the AskMSR System Note: these viewgraphs were originally developed by Professor Nick Kushmerick, University College Dublin,
The Vector Space Model …and applications in Information Retrieval.
CSCI 5582 Fall 2006 CSCI 5582 Artificial Intelligence Lecture 26 Jim Martin.
Information Retrieval
WMES3103 : INFORMATION RETRIEVAL INDEXING AND SEARCHING.
1 Question Answering, Information Retrieval, and Summarization Chapter 23 November 2012 Lecture #14 Most of this from Kathy McKeown (I left her format.
Chapter 5: Information Retrieval and Web Search
Overview of Search Engines
CSCI 5417 Information Retrieval Systems Jim Martin Lecture 6 9/8/2011.
Challenges in Information Retrieval and Language Modeling Michael Shepherd Dalhousie University Halifax, NS Canada.
MediaEval Workshop 2011 Pisa, Italy 1-2 September 2011.
COMP423: Intelligent Agent Text Representation. Menu – Bag of words – Phrase – Semantics – Bag of concepts – Semantic distance between two words.
A Technical Seminar on Question Answering SHRI RAMDEOBABA COLLEGE OF ENGINEERING & MANAGEMENT Presented By: Rohini Kamdi Guided By: Dr. A.J.Agrawal.
1 Chapter 19: Information Retrieval Chapter 19: Information Retrieval Relevance Ranking Using Terms Relevance Using Hyperlinks Synonyms., Homonyms,
Chapter 2 Architecture of a Search Engine. Search Engine Architecture n A software architecture consists of software components, the interfaces provided.
Question Answering.  Goal  Automatically answer questions submitted by humans in a natural language form  Approaches  Rely on techniques from diverse.
CSE 6331 © Leonidas Fegaras Information Retrieval 1 Information Retrieval and Web Search Engines Leonidas Fegaras.
Question Answering From Zero to Hero Elena Eneva 11 Oct 2001 Advanced IR Seminar.
A Probabilistic Graphical Model for Joint Answer Ranking in Question Answering Jeongwoo Ko, Luo Si, Eric Nyberg (SIGIR ’ 07) Speaker: Cho, Chin Wei Advisor:
Term Frequency. Term frequency Two factors: – A term that appears just once in a document is probably not as significant as a term that appears a number.
Chapter 6: Information Retrieval and Web Search
Search Engines. Search Strategies Define the search topic(s) and break it down into its component parts What terms, words or phrases do you use to describe.
Introduction to Digital Libraries hussein suleman uct cs honours 2003.
Comparing and Ranking Documents Once our search engine has retrieved a set of documents, we may want to Rank them by relevance –Which are the best fit.
Contextual Ranking of Keywords Using Click Data Utku Irmak, Vadim von Brzeski, Reiner Kraft Yahoo! Inc ICDE 09’ Datamining session Summarized.
Collocations and Information Management Applications Gregor Erbach Saarland University Saarbrücken.
Personalization with user’s local data Personalizing Search via Automated Analysis of Interests and Activities 1 Sungjick Lee Department of Electrical.
Vector Space Models.
1 Data Mining: Text Mining. 2 Information Retrieval Techniques Index Terms (Attribute) Selection: Stop list Word stem Index terms weighting methods Terms.
Comparing Document Segmentation for Passage Retrieval in Question Answering Jorg Tiedemann University of Groningen presented by: Moy’awiah Al-Shannaq
Mining Dependency Relations for Query Expansion in Passage Retrieval Renxu Sun, Chai-Huat Ong, Tat-Seng Chua National University of Singapore SIGIR2006.
Answer Mining by Combining Extraction Techniques with Abductive Reasoning Sanda Harabagiu, Dan Moldovan, Christine Clark, Mitchell Bowden, Jown Williams.
1 Answering English Questions by Computer Jim Martin University of Colorado Computer Science.
(Pseudo)-Relevance Feedback & Passage Retrieval Ling573 NLP Systems & Applications April 28, 2011.
1 Question Answering and Logistics. 2 Class Logistics  Comments on proposals will be returned next week and may be available as early as Monday  Look.
Xiaoying Gao Computer Science Victoria University of Wellington COMP307 NLP 4 Information Retrieval.
Feature Assignment LBSC 878 February 22, 1999 Douglas W. Oard and Dagobert Soergel.
Question Answering Passage Retrieval Using Dependency Relations (SIGIR 2005) (National University of Singapore) Hang Cui, Renxu Sun, Keya Li, Min-Yen Kan,
Information Retrieval and Web Search IR models: Vector Space Model Term Weighting Approaches Instructor: Rada Mihalcea.
CS791 - Technologies of Google Spring A Web­based Kernel Function for Measuring the Similarity of Short Text Snippets By Mehran Sahami, Timothy.
1 Text Categorization  Assigning documents to a fixed set of categories  Applications:  Web pages  Recommending pages  Yahoo-like classification hierarchies.
IR 6 Scoring, term weighting and the vector space model.
Information Retrieval in Practice
CSCE 590 Web Scraping – Information Extraction II
Designing Cross-Language Information Retrieval System using various Techniques of Query Expansion and Indexing for Improved Performance  Hello everyone,
Text Based Information Retrieval
Information Retrieval
Information Retrieval and Web Search
Information Retrieval and Web Search
IST 516 Fall 2011 Dongwon Lee, Ph.D.
Information Retrieval and Web Search
Chapter 5: Information Retrieval and Web Search
CS246: Information Retrieval
Semantic Similarity Methods in WordNet and their Application to Information Retrieval on the Web Yizhe Ge.
Chapter 31: Information Retrieval
Information Retrieval and Web Design
Chapter 19: Information Retrieval
Presentation transcript:

Information Retrieval and Question-Answering Julia Hirschberg CS 4705

Today Information Retrieval Review of Methods TREC IR Tracks Question Answering Factoid Q/A A Sample System: UT Dallas (Harabagiu) A simpler alternative from MSR

Information Retrieval Basic assumption `Meanings’ of documents can be captured by analyzing (counting) the words they contain Bag of words approach `Documents’ can be web pages, news articles, passages in articles,…

Inverted Index Fundamental operation required Ability to map from words to documents in a collection of documents Approach: Create an inverted index is of words and the document ids of the documents that contain them Dog: 1,2,8,100,119,210,400 Dog: 1:4,7:11,13:15,17

Stop Lists and Stemming Used by all IR systems Stop List Frequent (function/closed-class) words not indexed (of, the, a …) Reduces size of inverted index with virtual no loss of search accuracy Stemming issues Are dog and dogs separate entries or are they collapsed to dog?

Phrasal Search Google et al allows users to perform phrasal searches, e.g. big red dog Hint: they don’t grep the collection Add locational information to the index dog: 1{104}, 2{10}, etc red: 1{103},… big: 1{102},… Phrasal searches can operate incrementally by piecing the phrases together

Ranked Retrieval Inverted index is just the start Given a query, find out how relevant all the documents in the collection are to that query

Ad Hoc Retrieval Task

Representation Represent documents and queries as bit vectors N word types in collection Representation of document consists of a 1 for each corresponding word type that occurs in the document Compare two docs or a query and a doc by summing bits they have in common

Term Weighting Which words are more important? Local weight How important is this term to the meaning of this document?  How often does it occur in the document? Term Frequency (tf)

Global weight Tf-idf weighting How well does this term discriminate among the documents in the collection?  How many documents does it appear in? Inverse Document Frequency (idf) N= number of documents; ni = number of documents with term I Tf-idf weighting Weight of term i in vector for doc j is product of frequency in j with log of inverse document frequency in collection

Vector Space Model

Cosine Similarity Normalize by document length

Ad Hoc Retrieval Given a user query q and a document collection D Find vectors of all documents in D that contain any of the terms in q  candidate documents C Convert the q to a vector using weighting scheme used to represent documents in D Compute cosine similarity between q’s vector and vectors of C documents Sort result and return

Advanced Issues in IR Query Expansion Typical queries very short Expand user query using an initial search and taking words from top N docs, using a thesaurus, using term clustering or WordNet to find synonyms…. Tasks beyond Ad Hoc query support Passage Retrieval, Multilingual IR, Speech IR, Summarization, Question Answering…

Question-Answering Systems Beyond retrieving relevant documents -- Do people want answers to particular questions? Three kinds of systems Finding answers in document collections Interfaces to relational databases Mixed initiative dialog systems What kinds of questions do people want to ask?

Factoid Questions

Typical Q/A Architecture

UT Dallas Q/A Systems Contains many components used by other systems More complex in interesting ways Most work completed by 2001 Documentation: Paşca and Harabagiu, High-Performance Question Answering from Large Text Collections, SIGIR’01. Paşca and Harabagiu, Answer Mining from Online Documents, ACL’01. Harabagiu, Paşca, Maiorano: Experiments with Open-Domain Textual Question Answering. COLING’00

UT Dallas System Architecture Extracts and ranks passages using surface-text techniques Captures the semantics of the question Selects keywords for PR Extracts and ranks answers using NL techniques Question Semantics Q Question Processing Passage Retrieval Answer Extraction A Keywords Passages WordNet WordNet Document Retrieval Parser Parser NER NER

Question Processing Two main tasks Question classification: Determine the type of the answer Query formulation: Extract keywords from the question and formulate a query

Answer Types Factoid questions… Who, where, when, how many… Answers fall into limited, fairly predictable set of categories Who questions will be answered by… Where questions will be answered by … Generally, systems select answer types from a set of Named Entities, augmented with other types that are relatively easy to extract

Answer Types Can Be More Complicated Who questions can have organizations or countries as answers Who sells the most hybrid cars? Who exports the most wheat? Which questions can have people as answers Which president went to war with Mexico?

Taxonomy of Answer Types Contains ~9000 concepts reflecting expected answer types Merges NEs with the WordNet hierarchy

Answer Type Detection Use combination of hand-crafted rules and supervised machine learning to determine the right answer type for a question But how do we make use of this answer type once we hypothesize it?

Query Formulation: Extract Terms from Query Questions approximated by sets of unrelated words (lexical terms) Similar to bag-of-word IR models Question (from TREC QA track) Lexical terms Q002: What was the monetary value of the Nobel Peace Prize in 1989? monetary, value, Nobel, Peace, Prize Q003: What does the Peugeot company manufacture? Peugeot, company, manufacture Q004: How much did Mercury spend on advertising in 1993? Mercury, spend, advertising, 1993 Q005: What is the name of the managing director of Apricot Computer? name, managing, director, Apricot, Computer

Passage Retrieval Q A Question Processing Passage Retrieval Answer Extracts and ranks passages using surface-text techniques Captures the semantics of the question Selects keywords for PR Extracts and ranks answers using NL techniques Question Semantics Q Question Processing Passage Retrieval Answer Extraction A Keywords Passages WordNet WordNet Document Retrieval Parser Parser NER NER

Passage Retrieval Loop Passage Extraction Extract passages that contain all selected keywords Passage size and start position dynamic Passage quality assessed and keywords adjusted accordingly In first iteration use first 6 keywords selected If number of passages found is lower than a threshold  query too strict  drop a keyword If number of passages found is higher than a threshold  query too relaxed  add a keyword

Scoring the Passages k1 k2 k3 k2 k1 k1 k2 k3 k2 k1 k1 k2 k3 k2 k1 Passages scored based on keyword windows E.g., if question contains keywords: {k1, k2, k3, k4}, and a passage matches k1 and k2 twice, k3 once, and k4 not at all, following windows built: Window 1 Window 2 k1 k2 k3 k2 k1 k1 k2 k3 k2 k1 Window 3 Window 4 k1 k2 k3 k2 k1 k1 k2 k3 k2 k1

Passage ordering performed using a sort that involves three scores: Number of words from question recognized in same sequence in window Number of words that separate the most distant keywords in the window Number of unmatched keywords in the window

Q A Answer Extraction Question Processing Passage Retrieval Answer Extracts and ranks passages using surface-text techniques Captures the semantics of the question Selects keywords for PR Extracts and ranks answers using NL techniques Question Semantics Q Question Processing Passage Retrieval Answer Extraction A Keywords Passages WordNet WordNet Document Retrieval Parser Parser NER NER

Ranking Candidate Answers Q066: Name the first private citizen to fly in space. Answer type: Person Text passage: “Among them was Christa McAuliffe, the first private citizen to fly in space. Karen Allen, best known for her starring role in “Raiders of the Lost Ark”, plays McAuliffe. Brian Kerwin is featured as shuttle pilot Mike Smith...”

Ranking Candidate Answers Q066: Name the first private citizen to fly in space. Answer type: Person Text passage: “Among them was Christa McAuliffe, the first private citizen to fly in space. Karen Allen, best known for her starring role in “Raiders of the Lost Ark”, plays McAuliffe. Brian Kerwin is featured as shuttle pilot Mike Smith...” Best candidate answer: Christa McAuliffe How is this determined?

Features Used in Answer Ranking Number of question terms matched in the answer passage Number of question terms matched in the same phrase as the candidate answer Number of question terms matched in the same sentence as the candidate answer Flag set to 1 if the candidate answer is followed by a punctuation sign Number of question terms matched, separated from the candidate answer by at most three words and one comma Number of terms occurring in the same order in the answer passage as in the question Average distance from candidate answer to question term matches SIGIR ‘01

How does this approach compare to IE-based Q/A? When was Barack Obama born? Where was George Bush born? What college did John McCain attend? When did John F Kennedy die?

An Online QA System http://tangra.si.umich.edu/clair/NSIR/html/nsir.cgi

Is Q/A Different on the Web? In TREC (and most commercial applications), retrieval is performed against a small closed collection of texts More noise on the Web and more diversity Different formats Different genres How likely are you to find the actual question you asked? How likely are you to find a declarative version of your question?

AskMSR Rewrite questions to turn them into statements and search for the statements Simple rewrite rules to rewrite original question into form of a statement Must detect answer type Do IR on statement Extract answers of right type based on frequency of occurrence

AskMSR Example

Question-Rewriting Intuition: User’s question often syntactically close to sentences containing the answer Where is the Louvre Museum located? The Louvre Museum is located in Paris Who created the character of Scrooge? Charles Dickens created the character of Scrooge

Question Classification Classify question into one of seven categories Who is/was/are/were…? When is/did/will/are/were …? Where is/are/were …? Hand-crafted category-specific transformation rules e.g.: For where questions, move ‘is’ to all possible locations Look to the right of the query terms for the answer. “Where is the Louvre Museum located?”  “is the Louvre Museum located”  “the is Louvre Museum located”  “the Louvre is Museum located”  “the Louvre Museum is located”  “the Louvre Museum located is”

Query the Search Engine Send all rewrites to Web search engine Retrieve top N answers (100-200) For speed, rely just on search engine’s snippets, not full text of the actual document

Gather Ngrams Enumerate all Ngrams (N=1,2,3) in all retrieved snippets Weight of ngrams: occurrence count, each weighted by reliability (weight) of rewrite rule that fetched the document Example: “Who created the character of Scrooge?” Dickens 117 Christmas Carol 78 Charles Dickens 75 Disney 72 Carl Banks 54 A Christmas 41 Christmas Carol 45 Uncle 31

Filter Ngrams Each question type associated with one or more data-type filters (regular expressions for answer types) Boost score of ngrams that match expected answer type Lower score of ngrams that don’t match E.g. Filter for how-many queries prefers a number How many dogs pull a sled in the Iditarod? So… disprefer candidate ngrams like Dog race, run, Alaskan, dog racing Prefer candidiate ngrams like Pool (of)16 dogs

Tiling the Answers: Concatenate Overlaps Scores 20 15 10 merged, discard old n-grams Charles Dickens Dickens Mr Charles Start with highest scoring candidate and iterate using greedy approach Score 45 Mr Charles Dickens

Evaluation Usually based on TREC-devised metric In Q/A most frequent metric is Mean Reciprocal Rank Each system returns N answers Score is 1/<rank of first correct answer> Average score over all questions attempted

Results Standard TREC test-bed (TREC 2001) 1M documents; 900 questions AskMSR technique would have placed in top 9 of ~30 participants with MRR = 0.507 But….with access to Web…would have come in second on TREC 2001 Be suspicious of any after-the-bake-off is over results

Which Approach to Q/A is Better? Does it depend on question type? On document collection available? On? How can we handle harder questions, where answers are fluid and depend on putting together information from disparate texts over time? Who is Condoleezza Rice? Who is Stephen Harper? Why did San Francisco have to hand-count ballots in the last election?

Summary Information Retrieval Question Answering IE-based (e.g. Biadsy) UT Dallas style Web-based (e.g. AskMSR) Next: Summarization