Finding and Re-Finding Through Personalization Jaime Teevan MIT, CSAIL David Karger (advisor), Mark Ackerman, Sue Dumais, Rob Miller (committee), Eytan.

Slides:



Advertisements
Similar presentations
Haystack: Per-User Information Environment 1999 Conference on Information and Knowledge Management Eytan Adar et al Presented by Xiao Hu CS491CXZ.
Advertisements

1 Evaluation Rong Jin. 2 Evaluation  Evaluation is key to building effective and efficient search engines usually carried out in controlled experiments.
Bringing Order to the Web: Automatically Categorizing Search Results Hao Chen SIMS, UC Berkeley Susan Dumais Adaptive Systems & Interactions Microsoft.
FindAll: A Local Search Engine for Mobile Phones Aruna Balasubramanian University of Washington.
@ Carnegie Mellon Databases User-Centric Web Crawling Sandeep Pandey & Christopher Olston Carnegie Mellon University.
Personalization and Search Jaime Teevan Microsoft Research.
WSCD INTRODUCTION  Query suggestion has often been described as the process of making a user query resemble more closely the documents it is expected.
5/15/2015 copyright - All Rights Reserved 1 Home Educator’s Network, Inc. Present Training Module III Methods of Teaching.
1 Learning User Interaction Models for Predicting Web Search Result Preferences Eugene Agichtein Eric Brill Susan Dumais Robert Ragno Microsoft Research.
Zoetrope: Interacting with the Ephemeral Web Eytan Adar, Mira Dontcheva James Fogarty, Dan Weld University of Washington & Adobe Systems.
Personalizing Search via Automated Analysis of Interests and Activities Jaime Teevan Susan T.Dumains Eric Horvitz MIT,CSAILMicrosoft Researcher Microsoft.
Seesaw Personalized Web Search Jaime Teevan, MIT with Susan T. Dumais and Eric Horvitz, MSR.
Ryen W. White, Microsoft Research Jeff Huang, University of Washington.
Quality is about testing early and testing often Joe Apuzzo, Ngozi Nwana, Sweety Varghese Student/Faculty Research Day CSIS Pace University May 6th, 2005.
DiffIE: Changing How You View Changes on the Web DiffIE: Changing How You View Changes on the Web Jaime Teevan, Susan T. Dumais, Daniel J. Liebling, and.
Information Re-Retrieval: Repeat Queries in Yahoo’s Logs Jaime Teevan, Eytan Adar, Rosie Jones, Michael A. S. Potts SIGIR 2007.
Section 2: Finding and Refinding Jaime Teevan Microsoft Research 1.
Information Re-Retrieval Repeat Queries in Yahoo’s Logs Jaime Teevan (MSR), Eytan Adar (UW), Rosie Jones and Mike Potts (Yahoo) Presented by Hugo Zaragoza.
SLOW SEARCH Jaime Teevan, Kevyn Collins-Thompson, Ryen White, Susan Dumais and Yubin Kim.
AdWords Instructor: Dawn Rauscher. Quality Score in Action 0a2PVhPQhttp:// 0a2PVhPQ.
The Perfect Search Engine Is Not Enough Jaime Teevan †, Christine Alvarado †, Mark S. Ackerman ‡ and David R. Karger † † MIT, CSAIL ‡ University of Michigan.
Jaime Teevan Microsoft Research Finding and Re-Finding Personal Information.
Personalization of the Digital Library Experience: Progress and Prospects Nicholas J. Belkin Rutgers University, USA
The Confident Researcher: Google Away (Module 2) The Confident Researcher: Google Away 2.
Manipulating Attention in Computer Games Matthias Bernhard, Le Zhang, Michael Wimmer Institute of Computer Graphics and Algorithms Vienna University of.
Accessing Digital Memory Yahoo! Search Pad Vivian Lin Dufour 1 Oct 2009.
Facets of Personalization Jaime Teevan Microsoft Research (CLUES) with S. Dumais, E. Horvitz, D. Liebling, E. Adar, J. Elsas, R. Hughes.
Searching the Web Dr. Frank McCown Intro to Web Science Harding University This work is licensed under Creative Commons Attribution-NonCommercial 3.0Attribution-NonCommercial.
Intelligent Tutoring System for CS-I and II Laboratory Middle Tennessee State University J. Yoo, C. Pettey, S. Yoo J. Hankins, C. Li, S. Seo Supported.
1 Can People Collaborate to Improve the relevance of Search Results? Florian Eiteljörge June 11, 2013Florian Eiteljörge.
 Search Engine Search Engine  Steps to Search for webpages pertaining to a specific information Steps to Search for webpages pertaining to a specific.
Exploring Online Social Activities for Adaptive Search Personalization CIKM’10 Advisor : Jia Ling, Koh Speaker : SHENG HONG, CHUNG.
WebMining Web Mining By- Pawan Singh Piyush Arora Pooja Mansharamani Pramod Singh Praveen Kumar 1.
Surviving the Information Explosion Christine Alvarado and Jaime Teevan.
Implicit Acquisition of Context for Personalization of Information Retrieval Systems Chang Liu, Nicholas J. Belkin School of Communication and Information.
“W HERE ’ D IT GO ?” Jaime Teevan Microsoft Research.
Basic Search Engine Optimization. What is SEO?  SEO is an abbreviation for search engine optimization.
Discovering and Using Groups to Improve Personalized Search Jaime Teevan, Merrie Morris, Steve Bush Microsoft Research.
Information Retrieval Effectiveness of Folksonomies on the World Wide Web P. Jason Morrison.
Individualized Knowledge Access David Karger Lynn Andrea Stein Mark Ackerman Ralph Swick.
Personalizing Search Jaime Teevan, MIT Susan T. Dumais, MSR and Eric Horvitz, MSR.
21/11/20151Gianluca Demartini Ranking Clusters for Web Search Gianluca Demartini Paul–Alexandru Chirita Ingo Brunkhorst Wolfgang Nejdl L3S Info Lunch Hannover,
1 Internet Research Third Edition Unit A Searching the Internet Effectively.
CiteSight: Contextual Citation Recommendation with Differential Search Avishay Livne 1, Vivek Gokuladas 2, Jaime Teevan 3, Susan Dumais 3, Eytan Adar 1.
Retroactive Answering of Search Queries Beverly Yang Glen Jeh.
COLLABORATIVE SEARCH TECHNIQUES Submitted By: Shikha Singla MIT-872-2K11 M.Tech(2 nd Sem) Information Technology.
Understanding and Predicting Personal Navigation.
Post-Ranking query suggestion by diversifying search Chao Wang.
SLOW SEARCH WITH PEOPLE Jaime Teevan, Microsoft In collaboration with Michael S. Bernstein, Kevyn Collins- Thompson, Susan T. Dumais,
By Pamela Drake SEARCH ENGINE OPTIMIZATION. WHAT IS SEO? Search engine optimization (SEO) is the process of affecting the visibility of a website or a.
THE WEB CHANGES EVERYTHING Jaime Teevan, Microsoft
 SEO Terms A few additional terms Search site: This Web site lets you search through some kind of index or directory of Web sites, or perhaps both an.
Personalizing Web Search Jaime Teevan, MIT with Susan T. Dumais and Eric Horvitz, MSR.
UOS Personalized Search Zhang Tao 장도. Zhang Tao Data Mining Contents Overview 1 The Outride Approach 2 The outride Personalized Search System 3 Testing.
ASSOCIATIVE BROWSING Evaluating 1 Jin Y. Kim / W. Bruce Croft / David Smith by Simulation.
Helping People Find Information Better Jaime Teevan, MIT with Christine Alvarado, Mark Ackerman and David Karger.
1 Personalizing Search via Automated Analysis of Interests and Activities Jaime Teevan, MIT Susan T. Dumais, Microsoft Eric Horvitz, Microsoft SIGIR 2005.
Jaime Teevan MIT, CSAIL The Re:Search Engine. “Pick a card, any card.”
Seesaw Personalized Web Search Jaime Teevan, MIT with Susan T. Dumais and Eric Horvitz, MSR.
Potential for Personalization Transactions on Computer-Human Interaction, 17(1), March 2010 Data Mining for Understanding User Needs Jaime Teevan, Susan.
The Web Changes Everything How Dynamic Content Affects the Way People Find Online Jaime Teevan Microsoft Research (CLUES) with S. Dumais, D. Liebling,
SLOW SEARCH WITH PEOPLE Jaime Teevan, Microsoft In collaboration with Collins-Thompson, White, Dumais, Kim, Jeong, Morris, Liebling,
Jaime Teevan MIT, CSAIL. HCI at MIT ♠HCI Seminar ♥Fridays at 1:30 pm ♥URL: ♥Announcement mailing list ♠Applying HCI to.
Information Storage and Retrieval Fall Lecture 1: Introduction and History.
Improving searches through community clustering of information
Simultaneous Support for Finding and Re-Finding
Amazing The Re:Search Engine Jaime Teevan MIT, CSAIL.
Consistency During Search Without Stagnation
The Perfect Search Engine Is Not Enough
CS246: Leveraging User Feedback
Presentation transcript:

Finding and Re-Finding Through Personalization Jaime Teevan MIT, CSAIL David Karger (advisor), Mark Ackerman, Sue Dumais, Rob Miller (committee), Eytan Adar, Christine Alvarado, Eric Horvitz, Rosie Jones, and Michael Potts

Thesis Overview Supporting Finding –How people find –Individual differences affect finding –Personalized finding tool Supporting Re-Finding –How people re-find –Finding and re-finding conflict –Personalized finding and re-finding tool

Old New

Thesis Overview Supporting Finding –How people find –How individuals find –Personalized finding tool Supporting Re-Finding –How people re-find –Finding and re-finding conflict –Personalized finding and re-finding tool

Supporting Re-Finding How people re-find –People repeat searches –Look for old and new Finding and re-finding conflict –Result changes cause problems Personalized finding and re-finding tool –Identify what is memorable –Merge in new information

Supporting Re-Finding How people find –People repeat searches –Look for old and new Finding and re-finding conflict –Result changes cause problems Personalized finding and re-finding tool –Identify what is memorable –Merge in new information Query log analysis Memorability study Re:Search Engine

Related Work How people re-find –Know a lot of meta-information [Dumais] –Follow known paths [Capra] Changes cause problems re-finding –Dynamic menus [Shneiderman] –Dynamic search result lists [White] Relevance relative to expectation [Joachims]

Query Log Analysis Previous log analysis studies –People re-visit Web pages [Greenberg] –Query logs: Sessions [Jones] Yahoo! log analysis –114 people over the course of a year –13,060 queries and their clicks Can we identify re-finding behavior? What happens when results change?

Re-Finding Common Repeat query Repeat click Unique click 40%86% 33% 87% 38% 26% of queries of repeat queries

Change Reduces Re-Finding Results change rank Change reduces probability of repeat click –No rank change: 88% chance –Rank change: 53% chance Why? –Gone? –Not seen? –New results are better?

Change Slows Re-Finding Look at time to click as proxy for Ease Rank change  slower repeat click –Compared with initial search to click –No rank change: Re-click is faster –Rank change: Re-click is slower Changes interfere with re-finding  ?

Old New

“Pick a card, any card.”

Case 1Case 2Case 3Case 4Case 5Case 6

Your Card is GONE!

People Forget a Lot

Change Blindness

Old New

We still need magic!

Memorability Study Participants issued self-selected query After an hour, asked to fill out a survey 129 people remembered something

Memorability a Function of Rank

Remembered Results Ranked High

Old New

result list 1 result list 2 … result list n Re:Search Engine Architecture User client Web browser MergeIndex of past queries Result cache Search engine User interaction cache queryresult list query 1 query 2 … query n score 1 score 2 … score n result list

Components of Re:Search Engine Index of Past Queries Result Cache User Interaction Cache Merge Algorithm Index of past queries query query 1 query 2 … query n score 1 score 2 … score n result list 1 result list 2 … result list n Result cache query 1 query 2 … query n User interaction cache result list 1 result list 2 … result list n Merge result list

Index of Past Queries Studied how queries differ –Log analysis –Survey of how people remember queries Unimportant: case, stop words, word order Likelihood of re-finding deceases with time Get the user to tell us if they are re-finding –Encourage recognition, not recall Index of past queries query query 1 query 2 … query n score 1 score 2 … score n

Merge Algorithm Benefit of New Information score –How likely new result is to be useful… –…In a particular rank Memorability score –How likely old result is to be remembered… –…In a particular rank Chose list maximizes memorability and benefit of new information result list 1 result list 2 … result list n Merge result list

Benefit of New Information Ideal: Use search engine score Approximation: Use rank Results that are ranked higher are more likely to be seen –Greatest benefit given to highly ranked results being ranked highly

Memorability Score How memorable is a result? How likely is it to be remembered at a particular rank?

Choose Best Possible List Consider every combination Include at least three old and three new Min-cost network flow problem … … … … m2m2 m1m1 m 10 b 10 b2b2 b1b1 s t Old New Slots

Old New

Evaluation Does merged list look unchanged? –List recognition study Does merging make re-finding easier? –List interaction study Is search experience improved overall? –Longitudinal study

List Interaction Study 42 participants Two sessions a day apart –12 tasks each session Tasks based on queries Queries selected based on log analysis –Session 1 –Session 2 Re-finding New-finding (“stomach flu”) (“Symptoms of stomach flu?”) (“What to expect at the ER?”)

List Interaction Study

New 1 New 2 New 3 New 4 New 5 New 6 Old 5 New 1 Old 1 Old 7 New 2 New 3 New 4 Old 4 New 5 New 6 Old New Experimental Conditions Six re-finding tasks –Original result list –Dumb merging –Intelligent merging Six new-finding tasks –New result list –Dumb merging –Intelligent merging

Old New Experimental Conditions Six re-finding tasks –Original result list –Dumb merging –Intelligent merging Six new-finding tasks –New result list –Dumb merging –Intelligent merging Old 1 Old 2 Old 4 New 1 New 2 New 3 New 4 New 5 New 6 Old 10 Old 1 Old 2 Old 4 Old 10

Measures Performance –Correct –Time Subjective –Task difficulty –Result quality

Experimental Conditions Six re-finding tasks –Original result list –Dumb merging –Intelligent merging Six new-finding tasks –New result list –Dumb merging –Intelligent merging Faster, fewer clicks, more correct answers, and easier! Similar to Session 1

Results: Re-Finding PerformanceOriginalDumbIntelligent % correct96% Time (seconds) 99%88%

Results: Re-Finding SubjectiveOriginalDumbIntelligent % correct99%88%96% Time (seconds) Task difficulty1.57 Result quality

Results: Re-Finding OriginalDumbIntelligent % correct99%88%96% Time (seconds) Task difficulty Result quality List same? Intelligent merging better than Dumb Almost as good as the Original list Similarity 60%76%

Results: New-Finding PerformanceNewDumbIntelligent % correct73%74%84% Time (seconds)

Results: New-Finding SubjectiveNewDumbIntelligent % correct73%74%84% Time (seconds) Task difficulty Result quality

Results: New-Finding NewDumbIntelligent % correct73%74%84% Time (seconds) Task difficulty Result quality List same? Knowledge re-use can help No difference between New and Intelligent Similarity 38%50%61%

Results: Summary Re-finding –Intelligent merging better than Dumb –Almost as good as the Original list New-finding –Knowledge re-use can help –No difference between New and Intelligent Intelligent merging best of both worlds

Conclusion How people re-find –People repeat searches –Look for old and new Finding and re-finding conflict –Result changes cause problems Personalized finding and re-finding tool –Identify what is memorable –Merge in new information

Future Work Improve and generalize model –More sophisticated measures of memorability –Other types of lists (inboxes, directory listings) Effectively use model –Highlight change as well as hide it Present change at the right time –This talk’s focus: what and how –What about when to display new information?

Thesis Overview Supporting Finding –How people find –How individuals find –Personalized finding tool Supporting Re-Finding –How people re-find –Finding and re-finding conflict –Personalized finding and re-finding tool

David Karger (advisor), Mark Ackerman, Sue Dumais, Rob Miller (committee), Eytan Adar, Christine Alvarado, Eric Horvitz, Rosie Jones, and Michael Potts Thank You! Jaime Teevan