Relevance Models and Answer Granularity for Question Answering W. Bruce Croft and James Allan CIIR University of Massachusetts, Amherst.

Slides:



Advertisements
Similar presentations
Information Retrieval and Organisation Chapter 12 Language Models for Information Retrieval Dell Zhang Birkbeck, University of London.
Advertisements

Metadata in Carrot II Current metadata –TF.IDF for both documents and collections –Full-text index –Metadata are transferred between different nodes Potential.
Chapter 5: Introduction to Information Retrieval
SEARCHING QUESTION AND ANSWER ARCHIVES Dr. Jiwoon Jeon Presented by CHARANYA VENKATESH KUMAR.
Information Retrieval in Practice
Information Retrieval Models: Probabilistic Models
Chapter 7 Retrieval Models.
Information Retrieval in Practice
Search Engines and Information Retrieval
IR Challenges and Language Modeling. IR Achievements Search engines  Meta-search  Cross-lingual search  Factoid question answering  Filtering Statistical.
Incorporating Language Modeling into the Inference Network Retrieval Framework Don Metzler.
A Markov Random Field Model for Term Dependencies Donald Metzler and W. Bruce Croft University of Massachusetts, Amherst Center for Intelligent Information.
Modern Information Retrieval
Modern Information Retrieval Chapter 2 Modeling. Can keywords be used to represent a document or a query? keywords as query and matching as query processing.
Information Retrieval in Practice
Question Answering using Language Modeling Some workshop-level thoughts.
 Manmatha MetaSearch R. Manmatha, Center for Intelligent Information Retrieval, Computer Science Department, University of Massachusetts, Amherst.
1 LM Approaches to Filtering Richard Schwartz, BBN LM/IR ARDA 2002 September 11-12, 2002 UMASS.
Language Modeling Approaches for Information Retrieval Rong Jin.
Overview of Search Engines
Information Retrieval in Practice
1 Probabilistic Language-Model Based Document Retrieval.
CS598CXZ Course Summary ChengXiang Zhai Department of Computer Science University of Illinois, Urbana-Champaign.
Challenges in Information Retrieval and Language Modeling Michael Shepherd Dalhousie University Halifax, NS Canada.
Search Engines and Information Retrieval Chapter 1.
Probabilistic Model for Definitional Question Answering Kyoung-Soo Han, Young-In Song, and Hae-Chang Rim Korea University SIGIR 2006.
Relevance Models for QA Project Update University of Massachusetts, Amherst AQUAINT meeting December, 2002 Bruce Croft and James Allan, PIs.
Lemur Application toolkit Kanishka P Pathak Bioinformatics CIS 595.
Chapter 2 Architecture of a Search Engine. Search Engine Architecture n A software architecture consists of software components, the interfaces provided.
Question Answering.  Goal  Automatically answer questions submitted by humans in a natural language form  Approaches  Rely on techniques from diverse.
A Markov Random Field Model for Term Dependencies Donald Metzler W. Bruce Croft Present by Chia-Hao Lee.
Modern Information Retrieval: A Brief Overview By Amit Singhal Ranjan Dash.
Retrieval Models for Question and Answer Archives Xiaobing Xue, Jiwoon Jeon, W. Bruce Croft Computer Science Department University of Massachusetts, Google,
Predicting Question Quality Bruce Croft and Stephen Cronen-Townsend University of Massachusetts Amherst.
1 Automatic Classification of Bookmarked Web Pages Chris Staff Second Talk February 2007.
Web Image Retrieval Re-Ranking with Relevance Model Wei-Hao Lin, Rong Jin, Alexander Hauptmann Language Technologies Institute School of Computer Science.
Probabilistic Models of Novel Document Rankings for Faceted Topic Retrieval Ben Cartrette and Praveen Chandar Dept. of Computer and Information Science.
Collocations and Information Management Applications Gregor Erbach Saarland University Saarbrücken.
Searching the web Enormous amount of information –In 1994, 100 thousand pages indexed –In 1997, 100 million pages indexed –In June, 2000, 500 million pages.
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
1 Thi Nhu Truong, ChengXiang Zhai Paul Ogilvie, Bill Jerome John Lafferty, Jamie Callan Carnegie Mellon University David Fisher, Fangfang Feng Victor Lavrenko.
Chapter 23: Probabilistic Language Models April 13, 2004.
Information Retrieval at NLC Jianfeng Gao NLC Group, Microsoft Research China.
How Do We Find Information?. Key Questions  What are we looking for?  How do we find it?  Why is it difficult? “A prudent question is one-half of wisdom”
For Monday Read chapter 24, sections 1-3 Homework: –Chapter 23, exercise 8.
A Language Modeling Approach to Information Retrieval 한 경 수  Introduction  Previous Work  Model Description  Empirical Results  Conclusions.
Carnegie Mellon Novelty and Redundancy Detection in Adaptive Filtering Yi Zhang, Jamie Callan, Thomas Minka Carnegie Mellon University {yiz, callan,
Jhu-hlt-2004 © n.j. belkin 1 Information Retrieval: A Quick Overview Nicholas J. Belkin
Relevance-Based Language Models Victor Lavrenko and W.Bruce Croft Department of Computer Science University of Massachusetts, Amherst, MA SIGIR 2001.
Comparing Document Segmentation for Passage Retrieval in Question Answering Jorg Tiedemann University of Groningen presented by: Moy’awiah Al-Shannaq
Mining Dependency Relations for Query Expansion in Passage Retrieval Renxu Sun, Chai-Huat Ong, Tat-Seng Chua National University of Singapore SIGIR2006.
Language Modeling Putting a curve to the bag of words Courtesy of Chris Jordan.
1 Evaluating High Accuracy Retrieval Techniques Chirag Shah,W. Bruce Croft Center for Intelligent Information Retrieval Department of Computer Science.
1 Adaptive Subjective Triggers for Opinionated Document Retrieval (WSDM 09’) Kazuhiro Seki, Kuniaki Uehara Date: 11/02/09 Speaker: Hsu, Yu-Wen Advisor:
CS798: Information Retrieval Charlie Clarke Information retrieval is concerned with representing, searching, and manipulating.
Indri at TREC 2004: UMass Terabyte Track Overview Don Metzler University of Massachusetts, Amherst.
Xiaoying Gao Computer Science Victoria University of Wellington COMP307 NLP 4 Information Retrieval.
PAIR project progress report Yi-Ting Chou Shui-Lung Chuang Xuanhui Wang.
A Study of Smoothing Methods for Language Models Applied to Ad Hoc Information Retrieval Chengxiang Zhai, John Lafferty School of Computer Science Carnegie.
Bayesian Extension to the Language Model for Ad Hoc Information Retrieval Hugo Zaragoza, Djoerd Hiemstra, Michael Tipping Microsoft Research Cambridge,
Information Retrieval in Practice
Information Retrieval in Practice
Topic Modeling for Short Texts with Auxiliary Word Embeddings
Information Organization: Overview
Search Engine Architecture
Information Retrieval (in Practice)
What is IR? In the 70’s and 80’s, much of the research focused on document retrieval In 90’s TREC reinforced the view that IR = document retrieval Document.
John Lafferty, Chengxiang Zhai School of Computer Science
Junghoo “John” Cho UCLA
Information Organization: Overview
Presentation transcript:

Relevance Models and Answer Granularity for Question Answering W. Bruce Croft and James Allan CIIR University of Massachusetts, Amherst

Issues Formal basis for QA Question modeling Answer granularity Semi-structured data Answer updating

Formal basis for QA Problem – QA systems tend to be ad hoc combinations of NLP, IR and other techniques  Performance can be fragile  Requires considerable knowledge engineering  Difficult to transition between question and answer types some questions can be answered with “facts”, others require more Process of determining probability of correct answer very similar to probabilistic IR models  i.e. satisfying information need  better answers  better IR

QA Assumptions? IR first, then QA Single answer vs. ranked list Answer is transformed text vs. collection of source texts  who does the inference?  analyst’s current environment Questions are questions vs. “queries”  experience at WESTLAW  spoken questions? Answers are text extracts vs. people

Basic Approach Use language models as basis for a QA system  build on recent success in IR  test limits in QA Try to capture important aspects of QA  question types  answer granularity  multiple answers  feedback Develop learning methodologies for QA Currently more Qs than As

LM for QA (QALM) View question text as being generated by a mixture of relevance model and question model  relevance model is related to topic of question  question model is related to form of question Question models are associated with answer models Answers are document text generated by a mixture of relevance model and answer model TREC-style QA corresponds to specific set of question and answer models Default question and answer models leads to usual IR process  e.g. “what are the causes of asthma?”

Basic Approach Question text Question Model Relevance Model Answer Model Answer texts Language Models Estimate models Rank answers by probability

Relevance Models: idea For every topic, there is an underlying Relevance Model R Queries and relevant documents are samples from R Simple case: R is a distribution over vocabulary (unigram)  estimation straightforward if we had examples of relevant docs  use the fact that query is a sample from R to estimate parameters  use word co-occurrence statistics Relevance Model Query Relevant Documents

Relevance Model: Estimation with no training data Relevance Model israeli q1q1 palestinian q2q2 raids q3q3 ??? w R

Relevance Model: Putting things together sample probabilities palestinian israeli raids ??? q1q1 q2q2 q3q3 w M M M Relevance Model

Retrieval using Relevance Models Probability Ranking Principle:  P(w|R) is estimated by P(w|Q) (Relevance Model)  P(w|N) is estimated by collection probabilities P(w) Document-likelihood vs. query-likelihood approach  query expansion vs. document smoothing  can be applied to multiple granularities  i.e. sentence, passage, document, multi-document

Question Models Language modeling (or other techniques) can be used to classify questions based on predefined categories  best features need for each category need to be determined What is the best process for estimating the relevance and question models? How are new question models trained?  how many question models are enough?

Answer Models and Granularity Answer models are associated with question models  Best features need to be determined What is best process for estimating probability of answer texts given answer and relevance models? How is answer granularity modeled?  Learn default granularity for question category and backoff to other granularities  Is there an optimal granularity for a particular question/database? How are answer models trained?  Relevance feedback approach a possibility

Semi Structured Data How can structure and metadata indicated by markup be used to improve QA effectiveness?  Examples – tables, passages in Web pages Answer models will include markup and metadata features  Could be viewed as an indexing problem – what is an answer passage? – contiguous text too limited  Construct “virtual documents” Current demo: QuasmQuasm

Answer Updating Answers can change or evolve over time Dealing with multiple answers and answers arriving in streams of information are similar problems  Novelty detection in TDT task Evidence combination and answer granularity will be important components of a solution Filtering thresholds also important

Evaluation The CMU/UMass LEMUR toolkit will be used for development TREC QA data used for initial tests of general approach Web-based data used for evaluation of semi-structured data Modified TDT environment will be used for answer updating experiments Others….