 Examine two basic sources for implicit relevance feedback on the segment level for search personalization. Eye tracking Display time.

Slides:



Advertisements
Similar presentations
Accurately Interpreting Clickthrough Data as Implicit Feedback Joachims, Granka, Pan, Hembrooke, Gay Paper Presentation: Vinay Goel 10/27/05.
Advertisements

Evaluating the Robustness of Learning from Implicit Feedback Filip Radlinski Thorsten Joachims Presentation by Dinesh Bhirud
Georg Buscher Georg Buscher, Andreas Dengel, Ludger van Elst German Research Center for AI (DFKI) Knowledge Management Department Kaiserslautern, Germany.
Introduction to Information Retrieval
Introduction to Information Retrieval
1 Evaluation Rong Jin. 2 Evaluation  Evaluation is key to building effective and efficient search engines usually carried out in controlled experiments.
Improved TF-IDF Ranker
Optimizing search engines using clickthrough data
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
WSCD INTRODUCTION  Query suggestion has often been described as the process of making a user query resemble more closely the documents it is expected.
Introduction to Information Retrieval (Part 2) By Evren Ermis.
Query Dependent Pseudo-Relevance Feedback based on Wikipedia SIGIR ‘09 Advisor: Dr. Koh Jia-Ling Speaker: Lin, Yi-Jhen Date: 2010/01/24 1.
1 Learning User Interaction Models for Predicting Web Search Result Preferences Eugene Agichtein Eric Brill Susan Dumais Robert Ragno Microsoft Research.
A Quality Focused Crawler for Health Information Tim Tang.
Evaluating Search Engine
Search Engines and Information Retrieval
Personalizing Search via Automated Analysis of Interests and Activities Jaime Teevan Susan T.Dumains Eric Horvitz MIT,CSAILMicrosoft Researcher Microsoft.
Query Operations: Automatic Local Analysis. Introduction Difficulty of formulating user queries –Insufficient knowledge of the collection –Insufficient.
Database Management Systems, R. Ramakrishnan1 Computing Relevance, Similarity: The Vector Space Model Chapter 27, Part B Based on Larson and Hearst’s slides.
Re-ranking Documents Segments To Improve Access To Relevant Content in Information Retrieval Gary Madden Applied Computational Linguistics Dublin City.
Recall: Query Reformulation Approaches 1. Relevance feedback based vector model (Rocchio …) probabilistic model (Robertson & Sparck Jones, Croft…) 2. Cluster.
1 CS 430 / INFO 430 Information Retrieval Lecture 24 Usability 2.
An investigation of query expansion terms Gheorghe Muresan Rutgers University, School of Communication, Information and Library Science 4 Huntington St.,
The Relevance Model  A distribution over terms, given information need I, (Lavrenko and Croft 2001). For term r, P(I) can be dropped w/o affecting the.
Personalized Ontologies for Web Search and Caching Susan Gauch Information and Telecommunications Technology Center Electrical Engineering and Computer.
Harini Sridharan Stephen Duraski.  Behavioral Targeting (BT) is a technique that uses a user’s web-browsing behavior to determine which ads to display.
Search Engines and Information Retrieval Chapter 1.
Georg Buscher German Research Center for Artificial Intelligence (DFKI) Knowledge Management Department Kaiserslautern, Germany SIGIR 07 Doctoral Consortium.
A Comparative Study of Search Result Diversification Methods Wei Zheng and Hui Fang University of Delaware, Newark DE 19716, USA
Reyyan Yeniterzi Weakly-Supervised Discovery of Named Entities Using Web Search Queries Marius Pasca Google CIKM 2007.
A Simple Unsupervised Query Categorizer for Web Search Engines Prashant Ullegaddi and Vasudeva Varma Search and Information Extraction Lab Language Technologies.
Improving Web Search Ranking by Incorporating User Behavior Information Eugene Agichtein Eric Brill Susan Dumais Microsoft Research.
Chapter 2 Architecture of a Search Engine. Search Engine Architecture n A software architecture consists of software components, the interfaces provided.
UOS 1 Ontology Based Personalized Search Zhang Tao The University of Seoul.
When Experts Agree: Using Non-Affiliated Experts To Rank Popular Topics Meital Aizen.
Implicit Acquisition of Context for Personalization of Information Retrieval Systems Chang Liu, Nicholas J. Belkin School of Communication and Information.
Probabilistic Query Expansion Using Query Logs Hang Cui Tianjin University, China Ji-Rong Wen Microsoft Research Asia, China Jian-Yun Nie University of.
Presenter: Lung-Hao Lee ( 李龍豪 ) January 7, 309.
Giorgos Giannopoulos (IMIS/”Athena” R.C and NTU Athens, Greece) Theodore Dalamagas (IMIS/”Athena” R.C., Greece) Timos Sellis (IMIS/”Athena” R.C and NTU.
Implicit User Feedback Hongning Wang Explicit relevance feedback 2 Updated query Feedback Judgments: d 1 + d 2 - d 3 + … d k -... Query User judgment.
Personalized Search Xiao Liu
CS 533 Information Retrieval Systems.  Introduction  Connectivity Analysis  Kleinberg’s Algorithm  Problems Encountered  Improved Connectivity Analysis.
1 Computing Relevance, Similarity: The Vector Space Model.
CPSC 404 Laks V.S. Lakshmanan1 Computing Relevance, Similarity: The Vector Space Model Chapter 27, Part B Based on Larson and Hearst’s slides at UC-Berkeley.
Alexey Kolosoff, Michael Bogatyrev 1 Tula State University Faculty of Cybernetics Laboratory of Information Systems.
IR Theory: Relevance Feedback. Relevance Feedback: Example  Initial Results Search Engine2.
Personalizing Search Jaime Teevan, MIT Susan T. Dumais, MSR and Eric Horvitz, MSR.
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
Personalizing Web Search using Long Term Browsing History Nicolaas Matthijs, Cambridge Filip Radlinski, Microsoft In Proceedings of WSDM
Personalization with user’s local data Personalizing Search via Automated Analysis of Interests and Activities 1 Sungjick Lee Department of Electrical.
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
Summarizing Encyclopedic Term Descriptions on the Web from Coling 2004 Atsushi Fujii and Tetsuya Ishikawa Graduate School of Library, Information and Media.
Date: 2012/11/29 Author: Chen Wang, Keping Bi, Yunhua Hu, Hang Li, Guihong Cao Source: WSDM’12 Advisor: Jia-ling, Koh Speaker: Shun-Chen, Cheng.
1 Adaptive Subjective Triggers for Opinionated Document Retrieval (WSDM 09’) Kazuhiro Seki, Kuniaki Uehara Date: 11/02/09 Speaker: Hsu, Yu-Wen Advisor:
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Predicting User Interests from Contextual Information R. W. White, P. Bailey, L. Chen Microsoft (SIGIR 2009) Presenter : Jae-won Lee.
Personalizing Web Search Jaime Teevan, MIT with Susan T. Dumais and Eric Horvitz, MSR.
UIC at TREC 2006: Blog Track Wei Zhang Clement Yu Department of Computer Science University of Illinois at Chicago.
CS791 - Technologies of Google Spring A Web­based Kernel Function for Measuring the Similarity of Short Text Snippets By Mehran Sahami, Timothy.
1 Personalizing Search via Automated Analysis of Interests and Activities Jaime Teevan, MIT Susan T. Dumais, Microsoft Eric Horvitz, Microsoft SIGIR 2005.
University Of Seoul Ubiquitous Sensor Network Lab Query Dependent Pseudo-Relevance Feedback based on Wikipedia 전자전기컴퓨터공학 부 USN 연구실 G
Sampath Jayarathna Cal Poly Pomona
Accurately Interpreting Clickthrough Data as Implicit Feedback
Evaluation of IR Systems
Preface to the special issue on context-aware recommender systems
Author: Kazunari Sugiyama, etc. (WWW2004)
Learning Literature Search Models from Citation Behavior
CS246: Leveraging User Feedback
Shailesh Kochhar and Adam Vogel CS 498CXZ, Spring 2006
INF 141: Information Retrieval
Presentation transcript:

 Examine two basic sources for implicit relevance feedback on the segment level for search personalization. Eye tracking Display time

 Personalization  User Interaction Data - click-through, display time, scrolling, mouse movements. - Eye tracker  Eye Tracker is So Expensive.

 Segment-level Display Time vs Segment-level Feedback  How do they compare to each other and to a pseudo relevance feedback baseline on the segment level?  How much can they improve the ranking of a large commercial Web search engine through re-ranking and query expansion?

 Feedback on the Segment Level : User actions suggesting specific document parts are more interesting of relevant.  Two basic input sources for generating implicit feedback 1. segment-level display time based on scrolling behavior 2. Reading detection based on eye tracking

 Behavior-based evidence vs Non Behavior-based pseudo Relevance feedback  The most characteristic terms from those document parts should then be used for query expansion and re-ranking purposes within the same search session.

 Display time on the segment level can very simply be measured by analyzing scrolling behavior.  Using Javascript-based observation component. every single line of text contained in the document  the display time was coarsened to the level of paragraphs.

 1. It concatenates all documents viewed by the user within one search in one large virtual context document d.  2. a filtered context document d p is generated containing only those parts of d that got a display time greater than the threshold t in seconds.  3. based on the filtered context document d p, the most characteristic terms are extracted according to their tf x idf scores.

 Same way as DsplTime(t), but incorporates negative feedback while computing term scores in the following way.

 The basic assumption of this method is that users stop reading if the information is irrelevant; otherwise they continue reading. Thus, terms contained in short coherently read parts of the text tend to get lower scores than terms in long coherently read parts.

 pseudo relevance feedback on the segment level is a method to fine relevant document parts without explicitly logging user behavior.  Therefore, it is well-suited as a baseline for our behavior-based methods.

1. Similar to the methods above, a virtual context document d is created by concatenating all documents viewed by the user within one search session. 2. d is split into a set of paragraphs. Using Lucene with a BM25 ranking function, the user query is applied to the set of paragraphs resulting in a ranking score for each paragraph.

3. Out of the entire set, those paragraphs are selected that achieved a ranking score not worse than half of the ranking score achieved by the best matching paragraph. 4. the selected paragraphs are merged into one virtual document d p which forms the basis for tf x itf - based term extraction, like in the method DsplTime(t) from above.

1. They were given a relatively broad information need. 2. They looked through 4 preselected documents containing information relevant to their information need. 3. They had to pose 3 different queries according to more specific (sub)information needs to a search engine. 4. For each query they had to judge all delivered result entries.

 Two Procedures; Re-Ranking Query Expansion

 Precision at K. P(K) measures the fraction of relevant documents within the top K results. Take all positive relevance labels in setting (+++, ++, +) as positive feedback and all negative labels as negative feedback. The position of relevant documents within the top K results does not influence the value of P(K).

 DCG at K. The Discounted Cumulative Gain r(j) : corresponds to the relevance score of the jth result entry. The ratings +++,++,+ lead to relevance scores of 3, 2, 1, respectively, whereas all negative ratings get a relevance score of 0.

 MAP (at K). Mean Average Precision Average precision concerning one result list is calculated as the mean of the p(K) values after each relevant document was retrieved. Compute MAP at K by delimitating the considered result lists to the first K entries, i.e., computing P(i) only for i<=K

 Segment-level display time can be as valuable as eye-tracking-based feedback, both concerning re-ranking of search result and query expansion.

 Feedback based on display time on the segment level is much coarser than feedback from eye tracking.  For re-ranking & query expansion it did works as well as eye-tracking-based feedback.

 segment-level display time yields comparable results as eye-tracking-based feedback.  Thus, it should be considered in future personalization systems as an inexpensive but precise method for implicit feedback.

Thank you =]