Why IR test collections are so bad Mark Sanderson University of Sheffield.

Slides:



Advertisements
Similar presentations
Recuperação de Informação B Cap. 10: User Interfaces and Visualization , , , November 29, 1999.
Advertisements

Introduction to Information Retrieval Introduction to Information Retrieval Lecture 7: Scoring and results assembly.
Improving your paper SUGGESTIONS FOR SUCCESS. Writing = Revising  Writing IS a process  This paper WILL take hard work to get a good grade (or even.
 How practices shape identity: An exploration of Transition for Undergraduate Psychology Students.
Query Chains: Learning to Rank from Implicit Feedback Paper Authors: Filip Radlinski Thorsten Joachims Presented By: Steven Carr.
UCLA : GSE&IS : Department of Information StudiesJF : 276lec1.ppt : 5/2/2015 : 1 I N F S I N F O R M A T I O N R E T R I E V A L S Y S T E M S Week.
Buy, Build, Automate: Why you should Buy Your Taxonomy Tom Reamy Chief Knowledge Architect KAPS Group Knowledge Architecture Professional Services
Search Results Need to be Diverse Mark Sanderson University of Sheffield.
Section Based Relevance Feedback Student: Nat Young Supervisor: Prof. Mark Sanderson.
Overview of Collaborative Information Retrieval (CIR) at FIRE 2012 Debasis Ganguly, Johannes Leveling, Gareth Jones School of Computing, CNGL, Dublin City.
IE 423 – Design of Decision Support Systems Introduction to Data Base Management Systems and MS Access.
Kalervo Järvelin – Issues in Context Modeling – ESF IRiX Workshop - Glasgow THEORETICAL ISSUES IN CONTEXT MODELLING Kalervo Järvelin
Precision and Recall.
1 Software Maintenance and Evolution CSSE 575: Session 4, Part 1 Software Maintenance – Big Issues served up, Side order of Reifer Steve Chenoweth Office.
Search Engines and Information Retrieval
T.Sharon - A.Frank 1 Internet Resources Discovery (IRD) Classic Information Retrieval (IR)
Query Operations: Automatic Local Analysis. Introduction Difficulty of formulating user queries –Insufficient knowledge of the collection –Insufficient.
Heuristic Evaluation. Sources for today’s lecture: Professor James Landay: stic-evaluation/heuristic-evaluation.ppt.
Information Retrieval Concerned with the: Representation of Storage of Organization of, and Access to Information items.
Individuals and interactions
Information Retrieval in Practice
Experimental Components for the Evaluation of Interactive Information Retrieval Systems Pia Borlund Dawn Filan 3/30/04 610:551.
Retrieval Evaluation. Introduction Evaluation of implementations in computer science often is in terms of time and space complexity. With large document.
Problem Addressed The Navigation –Aided Retrieval tries to provide navigational aided query processing. It claims that the conventional Information Retrieval.
An investigation of query expansion terms Gheorghe Muresan Rutgers University, School of Communication, Information and Library Science 4 Huntington St.,
ISP 433/633 Week 6 IR Evaluation. Why Evaluate? Determine if the system is desirable Make comparative assessments.
What is a prototype? A prototype is a small scale model of your larger product. Can be a physical object, or a simple software program. Many physical.
 Mark Sanderson, University of Sheffield Poliomyelitis and Post-Polio TREC topic 302.
MS Access: Database Concepts Instructor: Vicki Weidler.
Modern Retrieval Evaluations Hongning Wang
Search Engines and Information Retrieval Chapter 1.
Which mindset do you and your students have? Fixed mindset Believe that their basic qualities, e.g. intelligence or talent are fixed Spend their time documenting.
Evaluation Experiments and Experience from the Perspective of Interactive Information Retrieval Ross Wilkinson Mingfang Wu ICT Centre CSIRO, Australia.
The Cognitive Perspective in Information Science Research Anthony Hughes Kristina Spurgin.
Usability testing IS 403: User Interface Design Shaun Kane.
Thanks to Bill Arms, Marti Hearst Documents. Last time Size of information –Continues to grow IR an old field, goes back to the ‘40s IR iterative process.
Making the most of the World Wide Wonderland 22nd February 2005 An introduction to effective internet searching Lorraine Sperring Learning & Teaching Resources.
The Effect of Collection Organization and Query Locality on IR Performance 2003/07/28 Park,
IT253: Computer Organization
FIXEDGROWTH Avoid or fear challenges Give up when things are tough Make excuses Don’t try so you won’t fail Refuse to learn from mistakes Judge your.
BY JACK G. AND IAN H. NXT Function Robots Project.
GUIDED BY DR. A. J. AGRAWAL Search Engine By Chetan R. Rathod.
Ben Carterette Paul Clough Evangelos Kanoulas Mark Sanderson.
 A Case For Interaction: A Study of Interactive Information Retrieval Behavior and Effectiveness By Jürgen Koenemann and Nicholas J. Belkin (1996) John.
TOPIC 6 DATABASES CONTENT : 6.1 DATABASE DESIGN 6.2 NORMOLISATION AND DATA MODELING 6.3 METHODS AND TOOLS FOR ANALYZING AND IMPLEMENTING DATABASE DESIGN.
BY JACK GAMBLE AND IAN HERDEGEN NXT Function Robots Project.
Information Retrieval CSE 8337 Spring 2007 Introduction/Overview Some Material for these slides obtained from: Modern Information Retrieval by Ricardo.
Recuperação de Informação Cap. 01: Introdução 21 de Fevereiro de 1999 Berthier Ribeiro-Neto.
Finding a PersonBOS Finding a Person! Building an algorithm to search for existing people in a system Rahn Lieberman Manager Emdeon Corp (Emdeon.com)
Visualization in Text Information Retrieval Ben Houston Exocortex Technologies Zack Jacobson CAC.
The Cross Language Image Retrieval Track: ImageCLEF Breakout session discussion.
Michael Bendersky, W. Bruce Croft Dept. of Computer Science Univ. of Massachusetts Amherst Amherst, MA SIGIR
Chapter. 3: Retrieval Evaluation 1/2/2016Dr. Almetwally Mostafa 1.
 SEO Terms A few additional terms Search site: This Web site lets you search through some kind of index or directory of Web sites, or perhaps both an.
By Hannah M By Hannah M.   The design of a good website are nice colors, organized, and not over filled. They usually don’t have a complex scene. 
© 2015 albert-learning.com How to talk to your boss How to talk to your boss!!
We are Gathered Here Together: Why? Introduction to the NII Workshop on Whole-Session Evaluation of Interactive Information Retrieval Systems Whole-Session.
Relevant Document Distribution Estimation Method for Resource Selection Luo Si and Jamie Callan School of Computer Science Carnegie Mellon University
WHIM- Spring ‘10 By:-Enza Desai. What is HCIR? Study of IR techniques that brings human intelligence into search process. Coined by Gary Marchionini.
Developments in Evaluation of Search Engines
Information Storage and Retrieval Fall Lecture 1: Introduction and History.
Dynamics GP SmartList and SmartList Designer
Modern Information Retrieval
Lecture 12: Relevance Feedback & Query Expansion - II
Thanks to Bill Arms, Marti Hearst
IR Theory: Evaluation Methods
Introduction to Information Retrieval
Recuperação de Informação
Structure of IR Systems
Introduction to Search Engines
Presentation transcript:

Why IR test collections are so bad Mark Sanderson University of Sheffield

What is the problem? Test collections (as a whole) (by & large) –Don’t support context Have a fixed context What do I mean by that? –Don’t support context of different data –Very restricted notion of user

Data Very few test collections out there –<100? Why? –Perception that test collections are hard to produce Implies people will use centrally built ones Don’t try to build their own

Users Test collections simulate –One time search User –Wants to enter a verbose query No spelling mistakes –Wants to search only once

Only once? We’ve long known “real search” is interactive –But generally thought that test collections still help in picking an algorithm for interactive search –I’m now not sure that this is true

If you only search once… …you want a retrieval system that –Uses pseudo-relevance feedback –Allows matching on a subset of query words –It doesn’t matter how long search takes –It doesn’t matter if the user understands how the search system did what it did. Cause they don’t need to re-formulate their query.

Why are test collections like this? Look back 30 years What was IR like? –Search intermediaries Expectations were that IR systems would replace intermediaries –Search once, look through lots of documents?

Test collections simulate “intermediary situation” rather well –Long query –Single search –User willing to look through long list of documents Can’t be bothered to re-formulate?

For common searching tasks If you are willing to search more than once –Do you want a search system tuned to single search?

Do we ditch the collections? No –Search engines use them a lot MSN Search –500 topic test collection per week

What do we do? We build more –Rapidly built and prototyping of test collections Helps with dynamic User modelling –Think about how to simulate people entering multiple searches in test collections

Conclusion I hope you agree test collections are bad But that there are potential ways of improving them.