1 Evaluation Rong Jin. 2 Evaluation  Evaluation is key to building effective and efficient search engines usually carried out in controlled experiments.

Slides:



Advertisements
Similar presentations
Google News Personalization: Scalable Online Collaborative Filtering
Advertisements

Accurately Interpreting Clickthrough Data as Implicit Feedback Joachims, Granka, Pan, Hembrooke, Gay Paper Presentation: Vinay Goel 10/27/05.
Introduction to Information Retrieval
Retrieval Evaluation J. H. Wang Mar. 18, Outline Chap. 3, Retrieval Evaluation –Retrieval Performance Evaluation –Reference Collections.
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
Site Level Noise Removal for Search Engines André Luiz da Costa Carvalho Federal University of Amazonas, Brazil Paul-Alexandru Chirita L3S and University.
1 Learning User Interaction Models for Predicting Web Search Result Preferences Eugene Agichtein Eric Brill Susan Dumais Robert Ragno Microsoft Research.
Evaluating Search Engine
Search Engines and Information Retrieval
Evaluation.  Allan, Ballesteros, Croft, and/or Turtle Types of Evaluation Might evaluate several aspects Evaluation generally comparative –System A vs.
Retrieval Evaluation. Brief Review Evaluation of implementations in computer science often is in terms of time and space complexity. With large document.
Reference Collections: Task Characteristics. TREC Collection Text REtrieval Conference (TREC) –sponsored by NIST and DARPA (1992-?) Comparing approaches.
Inferences About Process Quality
ISP 433/633 Week 6 IR Evaluation. Why Evaluate? Determine if the system is desirable Make comparative assessments.
Information Retrieval
Evaluation.  Allan, Ballesteros, Croft, and/or Turtle Types of Evaluation Might evaluate several aspects Evaluation generally comparative –System A vs.
The Relevance Model  A distribution over terms, given information need I, (Lavrenko and Croft 2001). For term r, P(I) can be dropped w/o affecting the.
Evaluation Information retrieval Web. Purposes of Evaluation System Performance Evaluation efficiency of data structures and methods operational profile.
Evaluation of Image Retrieval Results Relevant: images which meet user’s information need Irrelevant: images which don’t meet user’s information need Query:
Retrieval Evaluation Hongning Wang
Modern Retrieval Evaluations Hongning Wang
Evaluation David Kauchak cs458 Fall 2012 adapted from:
Evaluation David Kauchak cs160 Fall 2009 adapted from:
Search Engines and Information Retrieval Chapter 1.
Philosophy of IR Evaluation Ellen Voorhees. NIST Evaluation: How well does system meet information need? System evaluation: how good are document rankings?
IR Evaluation Evaluate what? –user satisfaction on specific task –speed –presentation (interface) issue –etc. My focus today: –comparative performance.
Searching the Web Dr. Frank McCown Intro to Web Science Harding University This work is licensed under Creative Commons Attribution-NonCommercial 3.0Attribution-NonCommercial.
Evaluating Search Engines in chapter 8 of the book Search Engines Information Retrieval in Practice Hongfei Yan.
UOS 1 Ontology Based Personalized Search Zhang Tao The University of Seoul.
Implicit User Feedback Hongning Wang Explicit relevance feedback 2 Updated query Feedback Judgments: d 1 + d 2 - d 3 + … d k -... Query User judgment.
IR System Evaluation Farhad Oroumchian. IR System Evaluation System-centered strategy –Given documents, queries, and relevance judgments –Try several.
Evaluation INST 734 Module 5 Doug Oard. Agenda Evaluation fundamentals Test collections: evaluating sets  Test collections: evaluating rankings Interleaving.
 Examine two basic sources for implicit relevance feedback on the segment level for search personalization. Eye tracking Display time.
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
ASSESSING LEARNING ALGORITHMS Yılmaz KILIÇASLAN. Assessing the performance of the learning algorithm A learning algorithm is good if it produces hypotheses.
A Repetition Based Measure for Verification of Text Collections and for Text Categorization Dmitry V.Khmelev Department of Mathematics, University of Toronto.
C.Watterscs64031 Evaluation Measures. C.Watterscs64032 Evaluation? Effectiveness? For whom? For what? Efficiency? Time? Computational Cost? Cost of missed.
Retrieval Evaluation Hongning Wang
Performance Measures. Why to Conduct Performance Evaluation? 2 n Evaluation is the key to building effective & efficient IR (information retrieval) systems.
Advantages of Query Biased Summaries in Information Retrieval by A. Tombros and M. Sanderson Presenters: Omer Erdil Albayrak Bilge Koroglu.
Post-Ranking query suggestion by diversifying search Chao Wang.
Chapter 8: Evaluation Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
Modern Retrieval Evaluations Hongning Wang
Evaluation of Information Retrieval Systems Xiangming Mu.
Evaluation. The major goal of IR is to search document relevant to a user query. The evaluation of the performance of IR systems relies on the notion.
Information Retrieval Quality of a Search Engine.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Learning to Rank: From Pairwise Approach to Listwise Approach Authors: Zhe Cao, Tao Qin, Tie-Yan Liu, Ming-Feng Tsai, and Hang Li Presenter: Davidson Date:
Information Retrieval Lecture 3 Introduction to Information Retrieval (Manning et al. 2007) Chapter 8 For the MSc Computer Science Programme Dell Zhang.
Introduction to Information Retrieval Introduction to Information Retrieval Lecture 10 Evaluation.
Navigation Aided Retrieval Shashank Pandit & Christopher Olston Carnegie Mellon & Yahoo.
Usefulness of Quality Click- through Data for Training Craig Macdonald, ladh Ounis Department of Computing Science University of Glasgow, Scotland, UK.
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008 Annotations by Michael L. Nelson.
INFORMATION RETRIEVAL MEASUREMENT OF RELEVANCE EFFECTIVENESS 1Adrienn Skrop.
Using Blog Properties to Improve Retrieval Gilad Mishne (ICWSM 2007)
Sampath Jayarathna Cal Poly Pomona
Evaluation Anisio Lacerda.
Evaluation of IR Systems
Lecture 10 Evaluation.
Evaluation.
Modern Information Retrieval
IR Theory: Evaluation Methods
Lecture 6 Evaluation.
Dr. Sampath Jayarathna Cal Poly Pomona
INF 141: Information Retrieval
Retrieval Evaluation - Measures
Retrieval Performance Evaluation - Measures
Retrieval Evaluation - Reference Collections
Dr. Sampath Jayarathna Cal Poly Pomona
Retrieval Evaluation - Reference Collections
Presentation transcript:

1 Evaluation Rong Jin

2 Evaluation  Evaluation is key to building effective and efficient search engines usually carried out in controlled experiments online testing can also be done  Effectiveness and efficiency are related High efficiency may be obtained at the price of effectiveness

3 Evaluation Corpus  Test collections consisting of documents, queries, and relevance judgments, e.g.,

4 Test Collections

5 TREC Topic Example

6 Relevance Judgments  Obtaining relevance judgments is an expensive, time-consuming process who does it? what are the instructions? what is the level of agreement?  TREC judgments depend on task being evaluated generally binary reasonable agreement because of “narrative”

7 Pooling  Exhaustive judgments for all documents in a collection is not practical  Pooling technique is used in TREC top k results (k varied between 50 and 200) from the rankings obtained by different search engines are merged into a pool duplicates are removed documents are presented in some random order to the relevance judges  Produces a large number of relevance judgments for each query, although still incomplete

8 Pooling Search engine 1 d1d2 d100d75 Search engine 2 d43d2 d501d75 Search engine 3 d1d13 d100d75 Query 1,000,000 docs d2 d43 d13 d75 d1 d100 d501 To be judged

9 Bias in Relevance Judgments  Relevance judgment is subjective  Disagreement among assessors

10 Combine Multiple Judgments  Judges disagree a lot. How to combine judgments from multiple reviewers ? Union Intersection Majority vote

11 Combine Multiple Judgments  Large impact on absolute performance numbers  Virtually no impact on ranking of systems

12 Query Logs  Used for tuning and evaluating search engines also for techniques such as query suggestion and spell checking  Typical contents User identifier or user session identifier Query terms - stored exactly as user entered List of URLs of results, their ranks on the result list, and whether they were clicked on Timestamp(s) - records the time of user events such as query submission, clicks

13 Query Logs  Clicks are not relevance judgments although they are correlated biased by a number of factors such as rank on result list  Can use clickthough data to predict preferences between pairs of documents appropriate for tasks with multiple levels of relevance, focused on user relevance various “policies” used to generate preferences

14 Example Click Policy  Skip Above and Skip Next click data generated preferences

15 Query Logs  Click data can also be aggregated to remove noise  Click distribution information can be used to identify clicks that have a higher frequency than would be expected high correlation with relevance e.g., using click deviation to filter clicks for preference-generation policies

16 Evaluation Metrics: Classification View Relevant Retrieved Irrelevant RetrievedIrrelevant Rejected Relevant Rejected Relevant Not relevant RetrievedNot Retrieved Doc Action

17 Evaluation Metrics: Example Recall = 2/6 = 0.33 Precision = 2/3 = 0.67

18 Evaluation Metrics: Example Recall = 5/6 = 0.83 Precision = 5/6 = 0.83

19 F Measure  Harmonic mean of recall and precision Why harmonic mean? harmonic mean emphasizes the importance of small values, whereas the arithmetic mean is affected more by outliers that are unusually large

20 Evaluation Metrics: Example Recall = 2/6 = 0.33 Precision = 2/3 = 0.67 F = 2*Recall*Precision/(Recall + Precision) = 2*0.33*0.67/( ) = 0.22

21 Evaluation Metrics: Example Recall = 5/6 = 0.83 Precision = 5/6 = 0.83 F = 2*Recall*Precision/(Recall + Precision) = 2*0.83*0.83/( ) = 0.83

22 Evaluation for Ranking  Average precision Averaging the precision values from the rank positions where a relevant document was retrieved Set precision values to be zero for the not retrieved documents

23 Average Precision: Example

24 Average Precision: Example

25 Average Precision: Example

26 Average Precision: Example Miss one relevant document

27 Average Precision: Example Miss two relevant documents

28 Mean Average Precision (MAP)

29 Mean Average Precision (MAP)  Summarize rankings from multiple queries by averaging average precision  Most commonly used measure in research papers  Assumes user is interested in finding many relevant documents for each query  Requires many relevance judgments in text collection

30 Recall-Precision Graph Multiple precision for some recalls

31 Interpolation where S is the set of observed (R,P) points  Defines precision at any recall level as the maximum precision observed in any recall- precision point at a higher recall level produces a step function defines precision at recall 0.0

Interpolation 32 Recall Interpolated Precision

Interpolation 33 Recall Interpolated Precision

Interpolation 34 Recall Interpolated Precision

Interpolation 35 Recall Interpolated Precision

Interpolation 36 Recall Interpolated Precision

Interpolation 37 Recall Interpolated Precision

Interpolation 38 Recall Interpolated Precision

39 Interpolation

40 Average Precision at Standard Recall Levels Only consider standard recall levels: varying from 0.0 to 1.0 at the incremental of 0.1 Recall-precision graph plotted by simply joining the average precision points at the standard recall levels

41 Average Recall-Precision Graph

42 Graph for 50 Queries

43 Focusing on Top Documents  Users tend to look at only the top part of the ranked result list to find relevant documents  Some search tasks have only one relevant document e.g., navigational search, question answering  Recall not appropriate instead need to measure how well the search engine does at retrieving relevant documents at very high ranks

44 Focusing on Top Documents  Precision at Rank R R typically 5, 10, 20 easy to compute, average, understand not sensitive to rank positions less than R  Reciprocal Rank reciprocal of the rank at which the first relevant document is retrieved Mean Reciprocal Rank (MRR) is the average of the reciprocal ranks over a set of queries very sensitive to rank position

45 MRR RR = 1/1 = 1 RR = 1/2 = 0.5 MRR = (1+0.5)/2 = 0.75

46 Discounted Cumulative Gain (DCG)  Popular measure for evaluating web search and related tasks Use graded relevance  Two assumptions: Highly relevant documents are more useful than marginally relevant document the lower the ranked position of a relevant document, the less useful it is for the user, since it is less likely to be examined

47 Discounted Cumulative Gain  Gain is accumulated starting at the top of the ranking and is discounted at lower ranks Typical discount is 1/log (rank) With base 2, the discount at rank 4 is 1/2, and at rank 8 it is 1/3

48 DCG Example  10 ranked documents judged on 0-3 relevance scale: 3, 2, 3, 0, 0, 1, 2, 2, 3, 0  discounted gain: 3, 2/1, 3/1.59, 0, 0, 1/2.59, 2/2.81, 2/3, 3/3.17, 0 = 3, 2, 1.89, 0, 0, 0.39, 0.71, 0.67, 0.95, 0  DCG: 3, 5, 6.89, 6.89, 6.89, 7.28, 7.99, 8.66, 9.61, 9.61

49 Efficiency Metrics  Query throughput The number of queries processed per second  Query latency The time between issuing a query and receiving a response, measured in millisecond Users consider instantaneous if the latency is less than 150 millisecond  Relation between query throughput and latency High throughput  handle multiple queries simultaneously  high latency

50 Significance Tests  Given the results from a number of queries, how can we conclude that ranking algorithm A is better than algorithm B?  A significance test null hypothesis: no difference between A and B alternative hypothesis: B is better than A the power of a test is the probability that the test will reject the null hypothesis correctly increasing the number of queries in the experiment also increases power of test

51 Example Experimental Results Significance level:  = 0.05 Probability for B=A

52 Example Experimental Results t-test t = 2.33 p-value = 0.02 Significance level:  = 0.05 Probability for B=A  B is better than A Probability for B=A is 0.02 Avg

53 Online Testing  Test (or even train) using live traffic on a search engine  Benefits: real users, less biased, large amounts of test data  Drawbacks: noisy data, can degrade user experience  Often done on small proportion (1-5%) of live traffic

54 Summary  No single measure is the correct one for any application choose measures appropriate for task use a combination shows different aspects of the system effectiveness  Use significance tests (t-test)  Analyze performance of individual queries