Presentation is loading. Please wait.

Presentation is loading. Please wait.

THE COMPLEX TASK OF MAKING SEARCH SIMPLE Jaime Teevan Microsoft Research UMAP 2015.

Similar presentations


Presentation on theme: "THE COMPLEX TASK OF MAKING SEARCH SIMPLE Jaime Teevan Microsoft Research UMAP 2015."— Presentation transcript:

1 THE COMPLEX TASK OF MAKING SEARCH SIMPLE Jaime Teevan (@jteevan) Microsoft Research UMAP 2015

2

3 THE WORLD WIDE WEB 20 YEARS AGO Content 2,700 websites (14%.com) Tools Mosaic only 1 year old Pre-Netscape, IE, Chrome 4 years pre-Google Search Engines 54,000 pages indexed by Lycos 1,500 queries per day

4 THE WORLD WIDE WEB TODAY Trillions of pages indexed. Billions of queries per day.

5 1996 We assume information is static. But web content changes!

6 SEARCH RESULTS CHANGE New, relevant content Improved ranking Personalization General instability Can change during a query!

7 SEARCH RESULTS CHANGE

8 BIGGEST CHANGE ON THE WEB Behavioral data.

9 It is impossible to separate a cube into two cubes, or a fourth power into two fourth powers, or in general, any power higher than the second, into two like powers. I have discovered a truly marvellous proof of this, which this margin is too narrow to contain. BEHAVIORAL DATA MANY YEARS AGO Marginalia adds value to books Students prefer annotated texts Do we lose marginalia when we move to digital documents? No! Scale makes it possible to look at experiences in the aggregate, and to tailor and personalize

10 PAST SURPRISES ABOUT WEB SEARCH Early log analysis  Excite logs from 1997, 1999  Silverstein et al. 1999; Jansen et al. 2000; Broder 2002 Queries are not 7 or 8 words long Advanced operators not used or “misused” Nobody used relevance feedback Lots of people search for sex Navigational behavior common Prior experience was with library search

11 SEARCH IS COMPLEX, MULTI-STEPPED PROCESS Typical query involves more than one click  59% of people return to search page after their first click  Clicked results often not the endpoint  People orienteer from results using context as a guide  Not all information needs can be expressed with current tools  Recognition is easier than recall Typical search session involves more than one query  40% of sessions contain multiple queries  Half of all search time spent in sessions of 30+ minutes Search tasks often involves more than one session  25% of queries are from multi-session tasks

12 IDENTIFYING VARIATION ACROSS INDIVIDUALS

13 WHICH QUERY HAS LESS VARIATION? campbells soup recipes v. vegetable soup recipe tiffany’s v. tiffany nytimes v. connecticut newspapers www.usajobs.gov v. federal government jobs singaporepools.com v. singapore pools

14 NAVIGATIONAL QUERIES WITH LOW VARIATION Use everyone’s clicks to identify queries with low click entropy  12% of the query volume  Only works for popular queries Clicks predicted only 72% of the time  Double the accuracy for the average query  But what is going on the other 28% of the time? Many typical navigational queries are not identified  People visit interior pages  craigslist – 3% visit http://geo.craigslist.org/iso/us/cahttp://geo.craigslist.org/iso/us/ca  People visit related pages  weather.com – 17% visit http://weather.yahoo.comhttp://weather.yahoo.com

15 INDIVIDUALS FOLLOW PATTERNS Getting ready in the morning. Getting to a webpage.

16

17 FINDING OFTEN INVOLVES REFINDING Repeat query (33%)  user modeling, adaptation, and personalization Repeat click (39%)  http://umap2015.com/ http://umap2015.com/  Query  umap Lots of repeats (43%) Repeat Query 33% New Query 67% Repeat Click New Click Repeat Query 33%29%4% New Query 67%10%57% 39%61%

18 IDENTIFYING PERSONAL NAVIGATION Use an individual’s clicks to identify repeat (query, click) pairs  15% of the query volume  Most occur fewer than 25 times in the logs Queries more ambiguous  Rarely contain a URL fragment  Click entropy the same as for general Web queries  Multiple meanings – enquirer  Found navigation – bed bugs  Serendipitous encounters – etsy National Enquirer Cincinnati Enquirer http://www.medicinenet.com/bed_bugs/article.htm [Informational] Etsy.com Regretsy.com (parody) 95%

19 SUPPORTING PERSONAL NAVIGATION Tom Bosley - Wikipedia, the free encyclopedia Thomas Edward "Tom" Bosley (October 1, 1927 October 19, 2010) was an American actor, best known for portraying Howard Cunningham on the long-running ABC sitcom Happy Days. Bosley was born in Chicago, the son of Dora and Benjamin Bosley. en.wikipedia.org/wiki/tom_bosley Tom Bosley - Wikipedia, the free encyclopedia Bosley died at 4:00 a.m. of heart failure on October 19, 2010, at a hospital near his home in Palm Springs, California. … His agent, Sheryl Abrams, said Bosley had been battling lung cancer. en.wikipedia.org/wiki/tom_bosley

20 PATTERNS A DOUBLE EDGED SWORD Patterns are predictable. Changing a pattern is confusing.

21 CHANGE INTERRUPTS PATTERNS Example: Dynamic menus  Put commonly used items at top  Slows menu item access Does search result change interfere with refinding?

22 CHANGE INTERRUPTS REFINDING When search result ordering changes people are  Less likely to click on a repeat result  Slower to click on a repeat result when they do  More likely to abandon their search Happens within a query and across sessions Even happens when the repeat result moves up! How to reconcile the benefits of change with the interruption?

23 USE MAGIC TO MINIMIZE INTERRUPTION

24 ABRACADABRA Magic happens.

25 YOUR CARD IS GONE!

26 CONSISTENCY ONLY MATTERS SOMETIMES

27 BIAS PERSONALIZATION BY EXPERIENCE

28 CREATE CHANGE BLIND WEB EXPERIENCES

29

30 THE COMPLEX TASK OF MAKING SEARCH SIMPLE Challenge: The web is complex  Tools change, content changes  Different people use the web differently Fortunately, individuals are simple  We are predictable, follow patterns  Predictability enables personalization Beware of breaking expectations!  Bias personalization by expectations  Create magic personal experiences

31 REFERENCES Broder. A taxonomy of web search. SIGIR Forum, 2002 Donato, Bonchi, Chi & Maarek. Do you want to take notes? Identifying research missions in Yahoo! Search Pad. WWW 2010. Dumais. Task-based search: A search engine perspective. NSF Task-Based Information Search Systems Workshop, 2013. Jansen, Spink & Saracevic. Real life, real users, and real needs: A study and analysis of user queries on the web. IP&M, 2000. Kim, Cramer, Teevan & Lagun. Understanding how people interact with web search results that change in real-time using implicit feedback. CIKM 2013. Lee, Teevan & de la Chica. Characterizing multi-click search behavior and the risks and opportunities of changing results during use. SIGIR 2014. Mitchell & Shneiderman. Dynamic versus static menus: An exploratory comparison. SIGCHI Bulletin, 1989. Selberg & Etzioni. On the instability of web search engines. RIAO 2000. Silverstein, Marais, Henzinger & Moricz. Analysis of a very large web search engine query log. SIGIR Forum, 1999. Somberg. A comparison of rule-based and positionally constant arrangements of computer menu items. CHI 1986. Svore, Teevan, Dumais & Kulkarni. Creating temporally dynamic web search snippets. SIGIR 2012. Teevan. The Re:Search Engine: Simultaneous support for finding and re- finding. UIST 2007. Teevan. How people recall, recognize and reuse search results. TOIS, 2008. Teevan, Alvarado, Ackerman & Karger. The perfect search engine is not enough: A study of orienteering behavior in directed search. CHI 2004. Teevan, Collins-Thompson, White & Dumais. Viewpoint: Slow search. CACM, 2014. Teevan, Collins-Thompson, White, Dumais & Kim. Slow search: Information retrieval without time constraints. HCIR 2013. Teevan, Cutrell, Fisher, Drucker, Ramos, Andrés & Hu. Visual snippets: Summarizing web pages for search and revisitation. CHI 2009. Teevan, Dumais & Horvitz. Potential for personalization. TOCHI, 2010. Teevan, Dumais & Liebling. To personalize or not to personalize: Modeling queries with variation in user intent. SIGIR 2008. Teevan, Liebling & Geetha. Understanding and predicting personal navigation. WSDM 2011. Tyler & Teevan. Large scale query log analysis of re-finding. WSDM 2010. More at: http://research.microsoft.com/~teevan/publications/http://research.microsoft.com/~teevan/publications/

32 THANK YOU! Jaime Teevan (@jteevan) teevan@microsoft.com

33 EXTRA SLIDES How search engines can make use of change to improve search.

34 CHANGE CAN IDENTIFY IMPORTANT TERMS Divergence from norm  cookbooks  frightfully  merrymaking  ingredient  latkes Staying power in page Time Sep. Oct. Nov. Dec.

35 CHANGE CAN IDENTIFY IMPORTANT SEGMENTS Page elements change at different rates Pages are revisited at different rates Resonance can serve as a filter for important content

36

37

38

39 EXTRA SLIDES Impact of change on refinding behavior.

40 Change to click  Unsatisfied initially  Gone > Down > Stay > Up  Satisfied initially  Stay > Down > Up > Gone Changes around click  Always benefit NSAT users  Best below the click for satisfied users NSATSAT Up 2.004.65 Stay 2.084.78 Down 2.204.75 Gone 2.314.61 NSATChangesStatic Above 2.302.21 Below 2.091.99 SATChangesStatic Above 4.93 Below 4.794.61 BUT CHANGE HELPS WITH FINDING!

41 EXTRA SLIDES Privacy issues and behavioral logs.

42 PUBLIC SOURCES OF BEHAVIORAL LOGS Public Web service content  Twitter, Facebook, Digg, Wikipedia Research efforts to create logs  Lemur Community Query Log Project  http://lemurstudy.cs.umass.edu/ http://lemurstudy.cs.umass.edu/  1 year of data collection = 6 seconds of Google logs Publicly released private logs  DonorsChoose.org  http://developer.donorschoose.org/the-data http://developer.donorschoose.org/the-data  Enron corpus, AOL search logs, Netflix ratings

43 EXAMPLE: AOL SEARCH DATASET August 4, 2006: Logs released to academic community  3 months, 650 thousand users, 20 million queries  Logs contain anonymized User IDs August 7, 2006: AOL pulled the files, but already mirrored August 9, 2006: New York Times identified Thelma Arnold  “A Face Is Exposed for AOL Searcher No. 4417749”  Queries for businesses, services in Lilburn, GA (pop. 11k)  Queries for Jarrett Arnold (and others of the Arnold clan)  NYT contacted all 14 people in Lilburn with Arnold surname  When contacted, Thelma Arnold acknowledged her queries August 21, 2006: 2 AOL employees fired, CTO resigned September, 2006: Class action lawsuit filed against AOL AnonIDQueryQueryTimeItemRankClickURL ----------------------------------------------------------- 1234567jitp 2006-04-04 18:18:181http://www.jitp.net/ 1234567jipt submission process2006-04-04 18:18:183http://www.jitp.net/m_mscript.php?p=2 1234567computational social scinece2006-04-24 09:19:32 1234567 computational social science2006-04-24 09:20:042http://socialcomplexity.gmu.edu/phd.php 1234567 seattle restaurants2006-04-24 09:25:502http://seattletimes.nwsource.com/rests 1234567 perlman montreal2006-04-24 10:15:144http://oldwww.acm.org/perlman/guide.html 1234567jitp 2006 notification2006-05-20 13:13:13 …

44 EXAMPLE: AOL SEARCH DATASET Other well known AOL users  User 927 how to kill your wife  User 711391 i love alaska  http://www.minimovies.org/documentaires/view/ilovealaska http://www.minimovies.org/documentaires/view/ilovealaska Anonymous IDs do not make logs anonymous  Contain directly identifiable information  Names, phone numbers, credit cards, social security numbers  Contain indirectly identifiable information  Example: Thelma’s queries  Birthdate, gender, zip code identifies 87% of Americans

45 EXAMPLE: NETFLIX CHALLENGE October 2, 2006: Netflix announces contest  Predict people’s ratings for a $1 million dollar prize  100 million ratings, 480k users, 17k movies  Very careful with anonymity post-AOL May 18, 2008: Data de-anonymized  Paper published by Narayanan & Shmatikov  Uses background knowledge from IMDB  Robust to perturbations in data December 17, 2009: Doe v. Netflix March 12, 2010: Netflix cancels second competition Ratings 1: [Movie 1 of 17770] 12, 3, 2006-04-18 [CustomerID, Rating, Date] 1234, 5, 2003-07-08 [CustomerID, Rating, Date] 2468, 1, 2005-11-12 [CustomerID, Rating, Date] … Movie Titles … 10120, 1982, “Bladerunner” 17690, 2007, “The Queen” … All customer identifying information has been removed; all that remains are ratings and dates. This follows our privacy policy... Even if, for example, you knew all your own ratings and their dates you probably couldn’t identify them reliably in the data because only a small sample was included (less than one tenth of our complete dataset) and that data was subject to perturbation.


Download ppt "THE COMPLEX TASK OF MAKING SEARCH SIMPLE Jaime Teevan Microsoft Research UMAP 2015."

Similar presentations


Ads by Google