SLOW SEARCH WITH PEOPLE Jaime Teevan, Microsoft In collaboration with Michael S. Bernstein, Kevyn Collins- Thompson, Susan T. Dumais,

Slides:



Advertisements
Similar presentations
Beliefs & Biases in Web Search
Advertisements

Microsoft® Office Access® 2007 Training
1 Evaluation Rong Jin. 2 Evaluation  Evaluation is key to building effective and efficient search engines usually carried out in controlled experiments.
Templates and Styles Excel Advanced. Templates are pre- designed and formatted spreadsheets –They provide consistency of layout/structure –They.
TAC Vista Security. Target  TAC Vista & Security Integration  Key customer groups –Existing TAC Vista users Provide features and hardware for security.
Evaluating Search Engine
Web Projections Learning from Contextual Subgraphs of the Web Jure Leskovec, CMU Susan Dumais, MSR Eric Horvitz, MSR.
Interfaces for Querying Collections. Information Retrieval Activities Selecting a collection –Lists, overviews, wizards, automatic selection Submitting.
Sigir’99 Inside Internet Search Engines: Search Jan Pedersen and William Chang.
DiffIE: Changing How You View Changes on the Web DiffIE: Changing How You View Changes on the Web Jaime Teevan, Susan T. Dumais, Daniel J. Liebling, and.
Online Learning for Web Query Generation: Finding Documents Matching a Minority Concept on the Web Rayid Ghani Accenture Technology Labs, USA Rosie Jones.
An investigation of query expansion terms Gheorghe Muresan Rutgers University, School of Communication, Information and Library Science 4 Huntington St.,
SEARCH ENGINES By, CH.KRISHNA MANOJ(Y5CS021), 3/4 B.TECH, VRSEC. 8/7/20151.
Section 2: Finding and Refinding Jaime Teevan Microsoft Research 1.
Live Meeting 2007: Presenter's companion Interact with attendees You’re using Live Meeting to conduct an orientation session for new hires, and now you.
Finding and Re-Finding Through Personalization Jaime Teevan MIT, CSAIL David Karger (advisor), Mark Ackerman, Sue Dumais, Rob Miller (committee), Eytan.
Information Re-Retrieval Repeat Queries in Yahoo’s Logs Jaime Teevan (MSR), Eytan Adar (UW), Rosie Jones and Mike Potts (Yahoo) Presented by Hugo Zaragoza.
SLOW SEARCH Jaime Teevan, Kevyn Collins-Thompson, Ryen White, Susan Dumais and Yubin Kim.
Social Network Question Asking. Sheila Brown Does anyone know if Keanu Reeves was born in Hawaii? Mon at 5:23pm ∙ Comment ∙ Like MethodToolUseAskerAskedSpeed.
TwitterSearch : A Comparison of Microblog Search and Web Search
Personalization of the Digital Library Experience: Progress and Prospects Nicholas J. Belkin Rutgers University, USA
What Do People Ask Social Networks? Meredith Ringel Morris, MSR Jaime Teevan, MSR Katrina Panovich, MIT.
Evaluation Experiments and Experience from the Perspective of Interactive Information Retrieval Ross Wilkinson Mingfang Wu ICT Centre CSIRO, Australia.
Facets of Personalization Jaime Teevan Microsoft Research (CLUES) with S. Dumais, E. Horvitz, D. Liebling, E. Adar, J. Elsas, R. Hughes.
Understanding and Predicting Graded Search Satisfaction Tang Yuk Yu 1.
1 Searching through the Internet Dr. Eslam Al Maghayreh Computer Science Department Yarmouk University.
1 Can People Collaborate to Improve the relevance of Search Results? Florian Eiteljörge June 11, 2013Florian Eiteljörge.
Microsoft Internet Explorer and the Internet Using Microsoft Explorer 5.
Mining the Web to Create Minority Language Corpora Rayid Ghani Accenture Technology Labs - Research Rosie Jones Carnegie Mellon University Dunja Mladenic.
Implicit Acquisition of Context for Personalization of Information Retrieval Systems Chang Liu, Nicholas J. Belkin School of Communication and Information.
1 Smart Data Online Using Offline Cardholder Maintenance October 2013 rbs.co.uk.
Detecting Dominant Locations from Search Queries Lee Wang, Chuang Wang, Xing Xie, Josh Forman, Yansheng Lu, Wei-Ying Ma, Ying Li SIGIR 2005.
Moodle (Course Management Systems). Forums, Chats, and Messaging.
Curtis Spencer Ezra Burgoyne An Internet Forum Index.
Question Answering over Implicitly Structured Web Content
Discovering and Using Groups to Improve Personalized Search Jaime Teevan, Merrie Morris, Steve Bush Microsoft Research.
Searching the web Enormous amount of information –In 1994, 100 thousand pages indexed –In 1997, 100 million pages indexed –In June, 2000, 500 million pages.
Personalizing Search Jaime Teevan, MIT Susan T. Dumais, MSR and Eric Horvitz, MSR.
Qi Guo Emory University Ryen White, Susan Dumais, Jue Wang, Blake Anderson Microsoft Presented by Tetsuya Sakai, Microsoft Research.
Adish Singla, Microsoft Bing Ryen W. White, Microsoft Research Jeff Huang, University of Washington.
CiteSight: Contextual Citation Recommendation with Differential Search Avishay Livne 1, Vivek Gokuladas 2, Jaime Teevan 3, Susan Dumais 3, Eytan Adar 1.
1 Lesson 18 Managing and Reporting Database Information Computer Literacy BASICS: A Comprehensive Guide to IC 3, 3 rd Edition Morrison / Wells.
Ranking Related Entities Components and Analyses CIKM’10 Advisor: Jia Ling, Koh Speaker: Yu Cheng, Hsieh.
Search Engine using Web Mining COMS E Web Enhanced Information Mgmt Prof. Gail Kaiser Presented By: Rupal Shah (UNI: rrs2146)
Understanding and Predicting Personal Navigation.
Authentication in the information industry: the challenges Rob Scaysbrook, International Sales Manager
Sigir’99 Inside Internet Search Engines: Spidering and Indexing Jan Pedersen and William Chang.
THE WEB CHANGES EVERYTHING Jaime Teevan, Microsoft
Navigation Aided Retrieval Shashank Pandit & Christopher Olston Carnegie Mellon & Yahoo.
Bringing Order to the Web : Automatically Categorizing Search Results Advisor : Dr. Hsu Graduate : Keng-Wei Chang Author : Hao Chen Susan Dumais.
1 A Methodology for automatic retrieval of similarly shaped machinable components Mark Ascher - Dept of ECE.
1 Personalizing Search via Automated Analysis of Interests and Activities Jaime Teevan, MIT Susan T. Dumais, Microsoft Eric Horvitz, Microsoft SIGIR 2005.
Jaime Teevan MIT, CSAIL The Re:Search Engine. “Pick a card, any card.”
Potential for Personalization Transactions on Computer-Human Interaction, 17(1), March 2010 Data Mining for Understanding User Needs Jaime Teevan, Susan.
SEARCH AND CONTEXT Susan Dumais, Microsoft Research INFO 320.
SLOW SEARCH WITH PEOPLE Jaime Teevan, Microsoft In collaboration with Collins-Thompson, White, Dumais, Kim, Jeong, Morris, Liebling,
SLOW SEARCH Jaime Teevan, Microsoft In collaboration with Michael S. Bernstein, Kevyn Collins- Thompson, Susan T. Dumais, Shamsi T.
SEARCH ENGINE by: by: B.Anudeep B.Anudeep Y5CS016 Y5CS016.
mi·cro·pro·duc·tiv·i·ty /ˈmīkrō prōˌdəkˈtivətē/ noun
Lesson 23 Managing and Reporting Database Information
Simultaneous Support for Finding and Re-Finding
Supporting Big Tasks through Microtasks
Amazing The Re:Search Engine Jaime Teevan MIT, CSAIL.
Supporting Collaborative Writing with Microtasks
Slow Search With People
Consistency During Search Without Stagnation
Struggling and Success in Web Search
CHI 2016 workshop on Microproductivity
University of Melbourne, Australia
mi·cro·pro·duc·tiv·i·ty /ˈmīkrō prōˌdəkˈtivətē/ noun
Presentation transcript:

SLOW SEARCH WITH PEOPLE Jaime Teevan, Microsoft In collaboration with Michael S. Bernstein, Kevyn Collins- Thompson, Susan T. Dumais, Shamsi T. Iqbal, Ece Kamar, Yubin Kim, Walter S. Lasecki, Daniel J. Liebling, Merrie Ringel Morris, Katrina Panovich, Ryen W. White, et al.

Slow Movements

Speed Focus in Search Reasonable

Not All Searches Need to Be Fast Long-term tasks Long search sessions Multi-session searches Social search Question asking Technologically limited Mobile devices Limited connectivity Search from space

Making Use of Additional Time

CROWDSOURCING Using human computation to improve search

Replace Components with People Search process Understand query Retrieve Understand results Machines are good at operating at scale People are good at understanding with Kim, Collins-Thompson

Understand Query: Query Expansion Original query: hubble telescope achievements Automatically identify expansion terms: space, star, astronomy, galaxy, solar, astro, earth, astronomer Best expansion terms cover multiple aspects of the query Ask crowd to relate expansion terms to a query term Identify best expansion terms: astronomer, astronomy, star spacestarastronomygalaxysolarastroearthastronomer hubble telescope achievements

Understand Results: Filtering Remove irrelevant results from list Ask crowd workers to vote on relevance Example: hubble telescope achievements

People Are Not Good Components Test corpora Difficult Web queries TREC Web Track queries Query expansion generally ineffective Query filtering Improves quality slightly Improves robustness Not worth the time and cost Need to use people in new ways

Understand Query: Identify Entities Search engines do poorly with long, complex queries Query: Italian restaurant in Squirrel Hill or Greenfield with a gluten-free menu and a fairly sophisticated atmosphere Crowd workers identify important attributes Given list of potential attributes Option add new attributes Example: cuisine, location, special diet, atmosphere Crowd workers match attributes to query Attributes used to issue a structured search with Kim, Collins-Thompson

Understand Results: Tabulate Crowd workers used to tabulate search results Given a query, result, attribute and value Does the result meet the attribute?

People Can Provide Rich Input Test corpus: Complex restaurant queries to Yelp Query understanding improves results Particularly for ambiguous or unconventional attributes Strong preference for the tabulated results People asked for additional columns (e.g., star rating) Those who liked the traditional results valued familiarity

Create Answers from Search Results Understand query Use log analysis to expand query to related queries Ask crowd if the query has an answer Retrieve: Identify a page with the answer via log analysis Understand results: Extract, format, and edit an answer with Bernstein, Dumais, Liebling, Horvitz

Community Answers with Bing Distill

Create Answers to Social Queries Understand query: Use crowd to identify questions Retrieve: Crowd generates a response Understand results: Vote on answers from crowd, friends with Jeong, Morris, Liebling

Working with an UNKNOWN CROWD Addressing the challenges of crowdsourcing search

Communicating with the Crowd How to tell the crowd what you are looking for? Trade off: Minimize the cost of giving information for the searcher Maximize the value of the information for the crowd with Salehi, Iqbal, Kamar

Guessing from Examples or Rating ? with Organisciak, Kalai, Dumais, Miller

Asking the Crowd to Guess v. Rate Guessing Requires fewer workers Fun for workers Hard to capture complex preferences Rating Requires many workers to find a good match Easy for workers Data reusable Rand.GuessRate Salt shakers Food (Boston) Food (Seattle) (RMSE for 5 workers)

Handwriting Imitation via “Rating” Task: Write Wizard’s Hex.

Handwriting Imitation via “Guessing” Task: Write Wizard’s Hex by imitating above text.

Extraction and Manipulation Threats with Lasecki, Kamar

Information Extraction Target task: Text recognition Attack task Complete target task Return answer from target: %32.8%

gun (36%), fun (26%), sun (12%) Task Manipulation Target task: Text recognition Attack task Enter “sun” as the answer for the attack task sun (75%)sun (28%)

Payment for Extraction Task

FRIENDSOURCING Using friends as a resource during the search process

Searching versus Asking

Friends respond quickly 58% of questions answered by the end of search Almost all answered by the end of the day Some answers confirmed search findings But many provided new information Information not available online Information not actively sought Social content with Morris, Panovich

Shaping the Replies from Friends Should I watch E.T.?

Shaping the Replies from Friends Larger networks provide better replies Faster replies in the morning, more in the evening Question phrasing important Include question mark Target the question at a group (even at anyone) Be brief (although context changes nature of replies) Early replies shape future replies Opportunity for friends and algorithms to collaborate to find the best content with Morris, Panovich

SELFSOURCING Supporting the information seeker as they search

Jumping to the Conclusion with Eickhoff, White, Dumais, André

Supporting Search through Structure Provide search recipes Understand query Retrieve Process results For specific task types For general search tasks Structure enables people to Complete harder tasks Search for complex things from their mobile devices Delegate parts of the task with Liebling, Lasecki

Algorithms + Experience

Algorithms + Experience = Confusion

Change Interrupts Finding When search result ordering changes people are Less likely to click on a repeat result Slower to click on a repeat result when they do More likely to abandon their search with Lee, de la Chica, Adar, Jones, Potts

Use Magic to Minimize Interruption

Abracadabra

Your Card is Gone!

Consistency Only Matters Sometimes

Bias Presentation by Experience

Make Slow Search Change Blind

Summary

Further Reading in Slow Search Slow Search Teevan, Collins-Thompson, White, Dumais. Viewpoint: Slow search. CACM Teevan, Collins-Thompson, White, Dumais, Kim. Slow search: Information retrieval without time constraints. HCIR Crowdsourcing Bernstein, Teevan, Dumais, Libeling, Horvitz. Direct answers for search queries in the long tail. CHI Jeong, Morris, Teevan, Liebling. A crowd-powered socially embedded search engine. ICWSM Kim, Collins-Thompson, Teevan. Using the crowd to improve search result ranking and the search experience. TIST (under review). Lasecki, Teevan, Kamar. Information extraction and manipulation threats in crowd-powered systems. CSCW Organisciak, Teevan, Dumais, Miller, Kalai. A crowd of your own: Crowdsourcing for on-demand personalization. HCOMP Salehi, Teevan, Iqbal, Kamar. Talking to the crowd: Communicating context in crowd work. CHI 2016 (under review). Friendsourcing Morris, Teevan, Panovich. A comparison of information seeking using search engines and social networks. ICWSM Morris, Teevan, Panovich. What do people ask their social networks, and why? A survey study of status message Q&A behavior. CHI Teevan, Morris, Panovich. Factors affecting response quantity, quality and speed in questions asked via online social networks. ICWSM Seflsourcing André, Teevan, Dumais. From x-rays to silly putty via Uranus: Serendipity and its role in web search. CHI Cheng, Teevan, Iqbal, Bernstein. Break it down: A comparison of macro- and microtasks. CHI Eickhoff, Teevan, White, Dumais. Lessons from the journey: A query log analysis of within-session learning. WSDM Lee, Teevan, de la Chica. Characterizing multi-click behavior and the risks and opportunities of changing results during use. SIGIR Teevan. How People Recall, recognize and reuse search results. TOIS Teevan, Adar, Jones, Potts. Information re-retrieval: Repeat queries in Yahoo's logs. SIGIR Teevan, Liebling, Lasecki. Selfsourcing personal tasks. CHI 2014.

QUESTIONS? Slow Search with People Jaime Teevan, Microsoft