Rutgers Information Interaction Lab at TREC 2005: Trying HARD N.J. Belkin, M. Cole, J. Gwizdka, Y.-L. Li, J.-J. Liu, G. Muresan, D. Roussinov*, C.A. Smith,

Slides:



Advertisements
Similar presentations
Critical Reading Strategies: Overview of Research Process
Advertisements

Enhance the Attractiveness of Studies in Science and Technology WP 6: Formal Hinders Kevin Kelly Trinity College Dublin WP 6 Co-ordinator.
Developing and Evaluating a Query Recommendation Feature to Assist Users with Online Information Seeking & Retrieval With graduate students: Karl Gyllstrom,
Overview of Collaborative Information Retrieval (CIR) at FIRE 2012 Debasis Ganguly, Johannes Leveling, Gareth Jones School of Computing, CNGL, Dublin City.
Query Dependent Pseudo-Relevance Feedback based on Wikipedia SIGIR ‘09 Advisor: Dr. Koh Jia-Ling Speaker: Lin, Yi-Jhen Date: 2010/01/24 1.
1 Learning User Interaction Models for Predicting Web Search Result Preferences Eugene Agichtein Eric Brill Susan Dumais Robert Ragno Microsoft Research.
Optimizing Estimated Loss Reduction for Active Sampling in Rank Learning Presented by Pinar Donmez joint work with Jaime G. Carbonell Language Technologies.
A Quality Focused Crawler for Health Information Tim Tang.
Information Retrieval in Practice
Project 4 U-Pick – A Project of Your Own Design Proposal Due: April 14 th (earlier ok) Project Due: April 25 th.
Rutgers’ HARD Track Experiences at TREC 2004 N.J. Belkin, I. Chaleva, M. Cole, Y.-L. Li, L. Liu, Y.-H. Liu, G. Muresan, C. L. Smith, Y. Sun, X.-J. Yuan,
Modern Information Retrieval
Retrieval Evaluation. Brief Review Evaluation of implementations in computer science often is in terms of time and space complexity. With large document.
Recall: Query Reformulation Approaches 1. Relevance feedback based vector model (Rocchio …) probabilistic model (Robertson & Sparck Jones, Croft…) 2. Cluster.
SIMS 202 Information Organization and Retrieval Prof. Marti Hearst and Prof. Ray Larson UC Berkeley SIMS Tues/Thurs 9:30-11:00am Fall 2000.
Queensland University of Technology An Ontology-based Mining Approach for User Search Intent Discovery Yan Shen, Yuefeng Li, Yue Xu, Renato Iannella, Abdulmohsen.
1 LM Approaches to Filtering Richard Schwartz, BBN LM/IR ARDA 2002 September 11-12, 2002 UMASS.
1 Automatic Identification of User Goals in Web Search Uichin Lee, Zhenyu Liu, Junghoo Cho Computer Science Department, UCLA {uclee, vicliu,
University of Kansas Department of Electrical Engineering and Computer Science Dr. Susan Gauch April 2005 I T T C Dr. Susan Gauch Personalized Search Based.
An investigation of query expansion terms Gheorghe Muresan Rutgers University, School of Communication, Information and Library Science 4 Huntington St.,
Personalizing the Digital Library Experience Nicholas J. Belkin, Jacek Gwizdka, Xiangmin Zhang SCILS, Rutgers University
Overview of Search Engines
An Experimental Comparison of Click Position-Bias Models Nick Craswell Onno Zoeter Michael Taylor Bill Ramsey Microsoft Research.
Personalization in Local Search Personalization of Content Ranking in the Context of Local Search Philip O’Brien, Xiao Luo, Tony Abou-Assaleh, Weizheng.
Personalization of the Digital Library Experience: Progress and Prospects Nicholas J. Belkin Rutgers University, USA
Minimal Test Collections for Retrieval Evaluation B. Carterette, J. Allan, R. Sitaraman University of Massachusetts Amherst SIGIR2006.
Evaluation Experiments and Experience from the Perspective of Interactive Information Retrieval Ross Wilkinson Mingfang Wu ICT Centre CSIRO, Australia.
Philosophy of IR Evaluation Ellen Voorhees. NIST Evaluation: How well does system meet information need? System evaluation: how good are document rankings?
IR Evaluation Evaluate what? –user satisfaction on specific task –speed –presentation (interface) issue –etc. My focus today: –comparative performance.
Redeeming Relevance for Subject Search in Citation Indexes Shannon Bradshaw The University of Iowa
Improving Web Search Ranking by Incorporating User Behavior Information Eugene Agichtein Eric Brill Susan Dumais Microsoft Research.
A Study on Query Expansion Methods for Patent Retrieval Walid MagdyGareth Jones Centre for Next Generation Localisation School of Computing Dublin City.
When Experts Agree: Using Non-Affiliated Experts To Rank Popular Topics Meital Aizen.
Query Expansion By: Sean McGettrick. What is Query Expansion? Query Expansion is the term given when a search engine adding search terms to a user’s weighted.
Estimating Topical Context by Diverging from External Resources SIGIR’13, July 28–August 1, 2013, Dublin, Ireland. Presenter: SHIH, KAI WUN Romain Deveaud.
Mining the Web to Create Minority Language Corpora Rayid Ghani Accenture Technology Labs - Research Rosie Jones Carnegie Mellon University Dunja Mladenic.
Implicit Acquisition of Context for Personalization of Information Retrieval Systems Chang Liu, Nicholas J. Belkin School of Communication and Information.
21/11/2002 The Integration of Lexical Knowledge and External Resources for QA Hui YANG, Tat-Seng Chua Pris, School of Computing.
Implicit User Feedback Hongning Wang Explicit relevance feedback 2 Updated query Feedback Judgments: d 1 + d 2 - d 3 + … d k -... Query User judgment.
Personalized Search Xiao Liu
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
Personalization with user’s local data Personalizing Search via Automated Analysis of Interests and Activities 1 Sungjick Lee Department of Electrical.
Qi Guo Emory University Ryen White, Susan Dumais, Jue Wang, Blake Anderson Microsoft Presented by Tetsuya Sakai, Microsoft Research.
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
Query Expansion By: Sean McGettrick. What is Query Expansion? Query Expansion is the term given when a search engine adding search terms to a user’s weighted.
Jhu-hlt-2004 © n.j. belkin 1 Information Retrieval: A Quick Overview Nicholas J. Belkin
Web Search and Text Mining Lecture 5. Outline Review of VSM More on LSI through SVD Term relatedness Probabilistic LSI.
Mining Dependency Relations for Query Expansion in Passage Retrieval Renxu Sun, Chai-Huat Ong, Tat-Seng Chua National University of Singapore SIGIR2006.
1 Evaluating High Accuracy Retrieval Techniques Chirag Shah,W. Bruce Croft Center for Intelligent Information Retrieval Department of Computer Science.
NTNU Speech Lab Dirichlet Mixtures for Query Estimation in Information Retrieval Mark D. Smucker, David Kulp, James Allan Center for Intelligent Information.
Generating Query Substitutions Alicia Wood. What is the problem to be solved?
WIDIT at TREC-2005 HARD Track Kiduk Yang, Ning Yu, Hui Zhang, Ivan Record, Shahrier Akram WIDIT Laboratory School of Library & Information Science Indiana.
Topic by Topic Performance of Information Retrieval Systems Walter Liggett National Institute of Standards and Technology TREC-7 (1999)
The Loquacious ( 愛說話 ) User: A Document-Independent Source of Terms for Query Expansion Diane Kelly et al. University of North Carolina at Chapel Hill.
Date: 2012/5/28 Source: Alexander Kotov. al(CIKM’11) Advisor: Jia-ling, Koh Speaker: Jiun Jia, Chiou Interactive Sense Feedback for Difficult Queries.
DISTRIBUTED INFORMATION RETRIEVAL Lee Won Hee.
Personalization Services in CADAL Zhang yin Zhuang Yuting Wu Jiangqin College of Computer Science, Zhejiang University November 19,2006.
Indri at TREC 2004: UMass Terabyte Track Overview Don Metzler University of Massachusetts, Amherst.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Toward Entity Retrieval over Structured and Text Data Mayssam Sayyadian, Azadeh Shakery, AnHai Doan, ChengXiang Zhai Department of Computer Science University.
ASSOCIATIVE BROWSING Evaluating 1 Jin Y. Kim / W. Bruce Croft / David Smith by Simulation.
Copyright © 2011 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 1 Research: An Overview.
Navigation Aided Retrieval Shashank Pandit & Christopher Olston Carnegie Mellon & Yahoo.
SIGIR 2005 Relevance Information: A Loss of Entropy but a Gain for IDF? Arjen P. de Vries Thomas Roelleke,
1 Personalizing Search via Automated Analysis of Interests and Activities Jaime Teevan, MIT Susan T. Dumais, Microsoft Eric Horvitz, Microsoft SIGIR 2005.
Using Blog Properties to Improve Retrieval Gilad Mishne (ICWSM 2007)
University Of Seoul Ubiquitous Sensor Network Lab Query Dependent Pseudo-Relevance Feedback based on Wikipedia 전자전기컴퓨터공학 부 USN 연구실 G
Writing a sound proposal
Lecture 12: Relevance Feedback & Query Expansion - II
Relevance and Reinforcement in Interactive Browsing
Presentation transcript:

Rutgers Information Interaction Lab at TREC 2005: Trying HARD N.J. Belkin, M. Cole, J. Gwizdka, Y.-L. Li, J.-J. Liu, G. Muresan, D. Roussinov*, C.A. Smith, A. Taylor, X.-J. Yuan Rutgers University; *Arizona State University

Our Major Goal Clarification forms (CFs) are simulations of user-system interaction Users are unwilling to engage in explicit interaction unless payoff is high, and interaction is understood as relevant Is explicit interaction worthwhile, and if so, under what circumstances?

General Approach to the Question Use relatively “standard” interactive elicitation techniques to enhance/ disambiguate original query Compare results to baseline Compare results to baseline plus relatively “standard” non-interactive query enhancement techniques, in particular, pseudo-rf

Methods for Automatic Query Enhancement Pseudo-relevance feedback (standard Lemur) Language modeling-based query expansion (clarity), derived from collection Web-based query expansion

Methods for User-Based Query Enhancement User selection of terms suggested by “clarity” and web methods (user selection based on Koenemann & Belkin, 1996; Belkin, et al., 2000) Elicitation of extended information problem descriptions (elicitation based on Kelly, Dollu & Fu, 2004; 2005)

Hypotheses for Automatic Enhancement H1: Query expansion using “clarity”- derived terms will improve performance over baseline & baseline + pseudo-rf H2: Query expansion using web-derived terms will improve performance, ditto H2b: Query expansion using both clarity- and web-derived terms will improve performance, ditto

Hypotheses for User-Based Query Enhancement H3: Query expansion with terms selected by the user from those suggested by clarity- and web-derived terms will improve performance, over everything else H4: Query expansion using “problem statements” elicited from users will increase performance over baseline & baseline + pseudo-rf

Hypothesis for When Elicitation is Useful H5: The effectiveness of query expansion using problem statements will be negatively correlated with query clarity.

Query Run Designations RUTGBL: Baseline query (title + description) RUTGBF3: Baseline + pseudo-rf (Lemur) RUTGWS1: Baseline + 0.1(Web-suggested terms) RUTGLS1: Baseline + 0.1(clarity-suggested terms) RUTGAS1: Baseline + 0.1(all suggested terms) RUTGUS1: Baseline + 0.1(terms selected by user) RUTGUG1: Baseline + 0.1(user-generated terms) RUTGALL: Baseline + all suggested terms and all user-generated terms

Identification of Suggested Terms Clarity: Compute query clarity for topic baseline (Lemur QueryClarity); sort terms accordingly; choose top ten Web: Next slide, please

Title: human smuggling Description: Identify incidents of human smuggling Navigation by Expansion Paradigm (NBE) aliens arrested border haitians trafficked undocumented

Navigation by Expansion Paradigm (NBE) Step1: Overview of the surroundings –Produces words and phrases “clearly related” to the topic –Internet mining: topic sent to Google –Logistic regression on the “signal to noise” ratio: Signal = df(results)/#results Noise = df(web)/#web Pr = 1 – exp (-(signal/noise – 1)/a) Step2: Valid “moves” identified –Related concepts from step 1 and those that Are present in AQUAINT Would affect search results if selected: impact estimate = P*df*idf Step 3: Selected moves executed –E.g. by query expansion: Score = original score + expansion score * expansion factor

“Combination” Run Combining pseudo-rf with user-selected terms from CF1 (run RUTBE) R-Prec. for RUTBE Substantially better than all other runs, but not comparable, because using different ranking function (BM25) and different differential weighting (0.3 for added terms) Indicative of possible improvements

User Selection (CF1)

User Generation (CF2)

System Implementation Lemur 3.1, 4.0, 4.1, using StructQueryEval Could we ask for somewhat more detailed documentation from the Lemur group?

Comparison to Other Sites R-precision MAP MeanSDMeanSDMeanSD Overall Baseline median RUTGBL Overall Final median RUTGALL0.299* **0.31

R-Precision for Test Runs

Summary of Significant Differences, R-Prec. BLAS1LS1WS1US1UG1BF3ALL BL AS1 * LS1 * n/s---- WS1 * n/s ---- US1 * n/s ---- UG1 * n/s ---- BF3 n/s n/s ---- ALL n/s n/s ----

Varying Weights of Baseline Terms w.r.t.CF2 Terms

Varying Weights of CF2 Terms w.r.t. Baseline Terms

CF2 & Baseline Terms, Equal Weights Run nameR-PrecisionPrecision at 10Mean Average Precision MeanSDMeanSDMeanSD RUTGBL Q * Q * Q ** **0.175 Q1Q20.298* ** **0.190 Q1Q30.313* *** **0.186 Q1Q2Q30.314** *** **0.190

Results w.r.t. Hypotheses H1, H2, H3, H4 weakly supported w.r.t. baseline, not to pseudo-rf H5 not supported –No correlation between baseline query clarity, and effectiveness of expanding with CF2 terms

Discussion (1) Both automatic and user-based query enhancement improved performance over baseline, but not over pseudo-rf No significant differences in performance between any enhancement methods, except Q1 v. Q1+Q3 (r-precision, vs )

Discussion (2) Some benefit both from automatic methods, and to explicit interaction with user, which require some effort from the user that goes beyond initial query formulation This interpretation of the results depends on the assumption that title+description queries are accurate simulations of user behavior

(Tentative) Conclusions Results indicate that invoking user interaction for query clarification is unlikely to be cost effective Alternative might be to develop ways to encourage more elaborate query formulation in the first instance, enhanced with automatic methods. Subsequent enhancement could be via implicit sources of evidence, rather than explicit questioning, requiring no additional effort from the user.