Accurately Interpreting Clickthrough Data as Implicit Feedback

Slides:



Advertisements
Similar presentations
Accurately Interpreting Clickthrough Data as Implicit Feedback Joachims, Granka, Pan, Hembrooke, Gay Paper Presentation: Vinay Goel 10/27/05.
Advertisements

Evaluating the Robustness of Learning from Implicit Feedback Filip Radlinski Thorsten Joachims Presentation by Dinesh Bhirud
Date: 2013/1/17 Author: Yang Liu, Ruihua Song, Yu Chen, Jian-Yun Nie and Ji-Rong Wen Source: SIGIR12 Advisor: Jia-ling Koh Speaker: Chen-Yu Huang Adaptive.
Introduction to Information Retrieval
1 Evaluation Rong Jin. 2 Evaluation  Evaluation is key to building effective and efficient search engines usually carried out in controlled experiments.
Modelling Relevance and User Behaviour in Sponsored Search using Click-Data Adarsh Prasad, IIT Delhi Advisors: Dinesh Govindaraj SVN Vishwanathan* Group:
Optimizing search engines using clickthrough data
Query Chains: Learning to Rank from Implicit Feedback Paper Authors: Filip Radlinski Thorsten Joachims Presented By: Steven Carr.
Eye Tracking Analysis of User Behavior in WWW Search Laura Granka Thorsten Joachims Geri Gay.
Web Mining Research: A Survey Authors: Raymond Kosala & Hendrik Blockeel Presenter: Ryan Patterson April 23rd 2014 CS332 Data Mining pg 01.
WSCD INTRODUCTION  Query suggestion has often been described as the process of making a user query resemble more closely the documents it is expected.
1 Learning User Interaction Models for Predicting Web Search Result Preferences Eugene Agichtein Eric Brill Susan Dumais Robert Ragno Microsoft Research.
Context-aware Query Suggestion by Mining Click-through and Session Data Authors: H. Cao et.al KDD 08 Presented by Shize Su 1.
1 Entity Ranking Using Wikipedia as a Pivot (CIKM 10’) Rianne Kaptein, Pavel Serdyukov, Arjen de Vries, Jaap Kamps 2010/12/14 Yu-wen,Hsu.
Evaluating Search Engine
Click Evidence Signals and Tasks Vishwa Vinay Microsoft Research, Cambridge.
1 CS 430 / INFO 430 Information Retrieval Lecture 8 Query Refinement: Relevance Feedback Information Filtering.
1 Discussion Class 11 Click through Data as Implicit Feedback.
Case study - usability evaluation Howell Istance.
1 CS 430 / INFO 430 Information Retrieval Lecture 24 Usability 2.
Problem Addressed The Navigation –Aided Retrieval tries to provide navigational aided query processing. It claims that the conventional Information Retrieval.
An investigation of query expansion terms Gheorghe Muresan Rutgers University, School of Communication, Information and Library Science 4 Huntington St.,
Evaluation CSC4170 Web Intelligence and Social Computing Tutorial 5 Tutor: Tom Chao Zhou
Advisor: Hsin-Hsi Chen Reporter: Chi-Hsin Yu Date:
Modern Retrieval Evaluations Hongning Wang
Evaluation David Kauchak cs458 Fall 2012 adapted from:
Evaluation David Kauchak cs160 Fall 2009 adapted from:
1 Context-Aware Search Personalization with Concept Preference CIKM’11 Advisor : Jia Ling, Koh Speaker : SHENG HONG, CHUNG.
Eye Tracking in the Design and Evaluation of Digital Libraries
Improving Web Search Ranking by Incorporating User Behavior Information Eugene Agichtein Eric Brill Susan Dumais Microsoft Research.
A Model of Information Foraging via Ant Colony Simulation Matthew Kusner.
WebMining Web Mining By- Pawan Singh Piyush Arora Pooja Mansharamani Pramod Singh Praveen Kumar 1.
Hao Wu Nov Outline Introduction Related Work Experiment Methods Results Conclusions & Next Steps.
Implicit Acquisition of Context for Personalization of Information Retrieval Systems Chang Liu, Nicholas J. Belkin School of Communication and Information.
Implicit User Feedback Hongning Wang Explicit relevance feedback 2 Updated query Feedback Judgments: d 1 + d 2 - d 3 + … d k -... Query User judgment.
Personalized Search Xiao Liu
Search Engines that Learn from Implicit Feedback Jiawen, Liu Speech Lab, CSIE National Taiwan Normal University Reference: Search Engines that Learn from.
 Examine two basic sources for implicit relevance feedback on the segment level for search personalization. Eye tracking Display time.
Qi Guo Emory University Ryen White, Susan Dumais, Jue Wang, Blake Anderson Microsoft Presented by Tetsuya Sakai, Microsoft Research.
AnnotatEd: A Social Navigation and Annotation Service for Web-based Educational Resources Rosta Farzan & Peter Brusilovsky Personalized Adaptive Web Systems.
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
Modeling Information Navigation : Implication for Information Architecture Craig S. Miller 이주우.
MOVIE RETRIEVAL SYSTEM INFORMATION VISUALIZATION & PROPOSING NEW INTERFACE IAT 814 Adrian Bisek.
Implicit User Feedback Hongning Wang Explicit relevance feedback 2 Updated query Feedback Judgments: d 1 + d 2 - d 3 + … d k -... Query User judgment.
Supporting Knowledge Discovery: Next Generation of Search Engines Qiaozhu Mei 04/21/2005.
Paired Experiments and Interleaving for Retrieval Evaluation Thorsten Joachims, Madhu Kurup, Filip Radlinski Department of Computer Science Department.
Active Feedback in Ad Hoc IR Xuehua Shen, ChengXiang Zhai Department of Computer Science University of Illinois, Urbana-Champaign.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
A Framework to Predict the Quality of Answers with Non-Textual Features Jiwoon Jeon, W. Bruce Croft(University of Massachusetts-Amherst) Joon Ho Lee (Soongsil.
Introduction to Information Retrieval Introduction to Information Retrieval Lecture 10 Evaluation.
1 CS 430 / INFO 430 Information Retrieval Lecture 12 Query Refinement and Relevance Feedback.
Introduction Web analysis includes the study of users’ behavior on the web Traffic analysis – Usage analysis Behavior at particular website or across.
Usefulness of Quality Click- through Data for Training Craig Macdonald, ladh Ounis Department of Computing Science University of Glasgow, Scotland, UK.
Potential for Personalization Transactions on Computer-Human Interaction, 17(1), March 2010 Data Mining for Understanding User Needs Jaime Teevan, Susan.
Evaluation Anisio Lacerda.
Search Engine Architecture
Lecture 12: Relevance Feedback & Query Expansion - II
Evaluation of IR Systems
Lesson 6: Databases and Web Search Engines
Content-Aware Click Modeling
A Study of Immediate Requery Behavior in Search
Mining Query Subtopics from Search Log Data
Beliefs and Biases in Web Search
Evidence from Behavior
Web Information retrieval (Web IR)
Date : 2013/1/10 Author : Lanbo Zhang, Yi Zhang, Yunfei Chen
CS246: Leveraging User Feedback
Cumulated Gain-Based Evaluation of IR Techniques
How does Clickthrough Data Reflect Retrieval Quality?
Interactive Information Retrieval
Presentation transcript:

Accurately Interpreting Clickthrough Data as Implicit Feedback Thorsten Joachims, Laura Granka, Bing Pan, Helene Hembrooke, & Geri Gay Cornell University SIGIR 2005 Presented by Rosta Farzan PAWS Group Meeting

Adapting retrieval systems requires large amount of data Problem Adapting retrieval systems requires large amount of data Implicit Data Explicit Data Expensive Noisy and unreliable

Goal Evaluate which types of implicit feedback can reliably be extracted from observed users behavior

Outline Introduction User Study Analysis Discussion

Introduction Designing a study to evaluate the reliability of implicit feedback How users interact with the list of ranked results from Google search Two types of analysis Analysis of users’ behavior Using eye-tracking & logging Do users scan from top to bottom? How many abstracts do they read before clicking? How does users’ behavior change if the result are manipulated artificially? Analysis of Implicit Feedback Comparing implicit feedback with explicit feedback collected manually

User Study Navigation Informational Task Participants Conditions Five navigational Find related web pages Five informational Find specific information Users read each question in turn and answered orally when they found the answer Participants Phase I 34 undergraduate, different major Used data from 29 because of eye-tracking issues Phase II 22 participants, 16 were used Conditions Normal - Google’s search result with no manipulation Swapped -Top two results were switched in order Reversed - 10 search results in reversed order Navigation Find the homepage of Michael Jordan, the statistician. Find the page displaying route map for Greyhound buses. Informational Where is the tallest mountain in New York located? Which actor starred as the main character in the original Time Machine movie?

User Study Data Collection Implicit data Explicit data HTTP-proxy server logs all click-stream data Eye-tracking fixations Explicit data Five judges for each two questions plus 10 results pages from two other questions Order the randomized results by how relevant they are Relative decision making Inter-judges agreement Phase I (ordering top 10): 89.5 % Phase II (ordering all results): 82.5%

Analysis of User Behavior Which links do users view and click? Do users scan links from top to bottom? Which links do users evaluate before clicking?

Which Links do Users View and Click? User click substantially more often on the first than second link Scrolling

Do Users Scan Links from Top to Bottom? On average users tend to read from top to bottom There is a big gap before viewing the third-ranked Users first scan the viewable results quite thoroughly before scrolling

Which Links do Users Evaluate before Clicking? They view substantially more abstracts above than below the click

Analysis of Implicit Feedback How relevance of the document to the query influence clicking decision? What Clicks tell us about the relevance of a document?

Does Relevance Influence User Decision? Using “reversed” condition Lower quality of retrieval Users react to the relevance of the presented links Users view lower ranked links more frequently Scan significantly more abstracts Users clicked less on first rank Users clicked more often on low ranked

Are Clicks Absolute Relevance Judgments? Trust bias Ranked first receives many more clicks Quality bias Comparing clicking behavior in “normal” condition vs. “reversed” condition. On lower quality, users click on abstracts that are on average less relevant

Are Clicks Relative Relevance Judgments? Consider not-clicked links as well as clicks as feedback signals Example: l1 l2 l3 l4 l5 l6 l7 Strategy 1 – Click > Skip Above Rel(l3) > rel(l2), rel(l5) > rel(l2), rel(l5) > rel(l4) Phase I data supports this strategy but phase II doesn’t Strategy 2 – Last Click > Skip Above Earlier clicks might be less informed than later clicks Rel(l5) > rel(l2), rel(l5) > rel(l4) Still not supported by phase II data

Strategies Strategy 3 – Click > Earlier Click Click later in time are on more relevant abstracts Assuming order of clicks as 3, 1, 5 Rel(l1)>rel(l3), rel(l5)>rel(l3), rel(l5)>rel(l1) Not supported by data Strategy 4 – Last Click > Skip Previous Constraint only between a clicked link and a not-clicked link immediately above Result is similar to strategy 1 Strategy 5 – Click > No-Click Next Constraint between a clicked link and an immediately following link