Presentation is loading. Please wait.

Presentation is loading. Please wait.

Observational Approaches to Information Retrieval SIGIR 2014 Tutorial: Choices and Constraints (Part II) Diane Kelly, Filip Radlinski, Jaime Teevan Slides.

Similar presentations


Presentation on theme: "Observational Approaches to Information Retrieval SIGIR 2014 Tutorial: Choices and Constraints (Part II) Diane Kelly, Filip Radlinski, Jaime Teevan Slides."— Presentation transcript:

1 Observational Approaches to Information Retrieval SIGIR 2014 Tutorial: Choices and Constraints (Part II) Diane Kelly, Filip Radlinski, Jaime Teevan Slides available at: http://aka.ms/sigirtutorial

2 Diane Kelly, University of North Carolina, USA Filip Radlinski, Microsoft, UK Jaime Teevan, Microsoft Research, USA

3 Tutorial Goals 1. To help participants develop a broader perspective of research goals and approaches in IR.  Descriptive, predictive and explanatory 2. To improve participants’ understandings of research choices and constraints.  Every research project requires the researcher to make a series of choices about a range of factors and usually there are constraints that influence these choices.  By using some of our own research papers, we aim to expose you to the experiential aspects of the research process by giving you a behind the scenes view of how we make/made choices in our own research.

4 Research Goals & Approaches Describe Report a set of observations and provide benchmarks (e.g., average queries per user, problems a user experiences when engaging in search) Such studies might also present categorizations of the observations Predict Seek to establish predictable relationships Take as input some set of features (click through rate, dwell time) and use these to predict other variables (query abandonment, satisfaction) Explain (why?) Propose a theoretical model that explains how select constructs interact and interrelate Devise procedures to measure those constructs (that is, translate the constructs into variables that can be controlled and measured) Devise protocol (usually experimental) to observe phenomenon of interest. Seek to demonstrate causality, not just show the variables are related.

5 Research Goals & Approaches DescribePredictExplain Afternoon Field Observation ✔ Log Analysis ✔✔ Morning Laboratory Experiment ✔✔✔ Field Experiment ✔✔✔

6 Example: Search Difficulties Describe A diary study might be used to gain insight about when and how users’ experience and address search difficulties. Log data might also be analyzed to identify how often these events occur. Predict A model might be constructed using the signals available in a log to predict when users will abandon search result pages without clicking. This model might then be evaluated with other log data. Explain Results from these studies might then be used to create an explanatory/theoretical model of search difficulty, which can be used to generate testable hypotheses. The model can include constructs and variables beyond those which are available in the log data. An experiment might be designed to test the explanatory power of the theory indirectly by examining the predictive power of the hypotheses.

7 Overview  Observational log analysis  What we can learn  Collecting log data  Cleaning log data (Filip)  Analyzing log data  Field observations (Diane) Dumais, Jeffries, Russell, Tang & Teevan. “Understanding User Behavior through Log Data and Analysis.”

8 What We Can Learn Observational Approaches to Information Retrieval

9 David Foster Wallace Mark Twain Cowards die many times before their deaths. Annotated by Nelson Mandela Cowards die many times before their deaths. Annotated by Nelson Mandela I have discovered a truly marvelous proof... which this margin is too narrow to contain. Pierre de Fermat (1637) Students prefer used textbooks that are annotated. [Marshall 1998]

10 Digital Marginalia  Do we lose marginalia with digital documents?  Internet exposes information experiences  Meta-data, annotations, relationships  Large-scale information usage data  Change in focus  With marginalia, interest is in the individual  Now we can look at experiences in the aggregate

11

12 Practical Uses for Behavioral Data  Behavioral data to improve Web search  Offline log analysis  Example: Re-finding common, so add history support  Online log-based experiments  Example: Interleave different rankings to find best algorithm  Log-based functionality  Example: Boost clicked results in a search result list  Behavioral data on the desktop  Goal: Allocate editorial resources to create Help docs  How to do so without knowing what people search for?

13 Value of Observational Log Analysis  Focus of observational log analysis  Description: What do people currently do?  Prediction: What will people do in similar situations?  Study real behavior in natural settings  Understand how people search  Identify real problems to study  Improve ranking algorithms  Influence system design  Create realistic simulations and evaluations  Build a picture of human interest

14 Societal Uses of Behavioral Data  Understand people’s information needs  Understand what people talk about  Impact public policy? (E.g., DonorsChoose.org) Baeza-Yates, Dupret, Velasco. A study of mobile search queries in Japan. WWW 2007

15 Personal Use of Behavioral Data  Individuals now have a lot of behavioral data  Introspection of personal data popular  My Year in Status  Status Statistics  Expect to see more  As compared to others  For a purpose

16 Defining Behavioral Log Data  Behavioral log data are:  Traces of natural behavior, seen through a sensor  Examples: Links clicked, queries issued, tweets posted  Real-world, large-scale, real-time  Behavioral log data are not:  Non-behavioral sources of large-scale data  Collected data (e.g., poll data, surveys, census data)  Not recalled behavior or subjective impression

17 Real-World, Large-Scale, Real-Time  Private behavior is exposed  Example: Porn queries, medical queries  Rare behavior is common  Example: Observe 500 million queries a day  Interested in behavior that occurs 0.002% of the time  Still observe the behavior 10 thousand times a day!  New behavior appears immediately  Example: Google Flu Trends

18 Drawbacks  Not controlled  Can run controlled log studies  Discussed in morning tutorial (Filip)  Adversarial  Cleaning log data later today (Filip)  Lots of missing information  Not annotated, no demographics, we don’t know why  Observing richer information after break (Diane)  Privacy concerns  Collect and store data thoughtfully  Next section addresses privacy

19 QueryTimeUser sigir 201410:41 am 1/15/14142039 goldcoast sofitel10:44 am 1/15/14142039 learning to rank10:56 am 1/15/14142039 sigir 201411:21 am 1/15/14659327 ool transportation11:59 am 1/15/14318222 restaurants brisbane12:01 pm 1/15/14318222 surf lessons12:17 pm 1/15/14318222 james allen12:18 pm 1/15/14142039 daytrips from brisbane1:30 pm 1/15/14554320 sigir 20141:30 pm 1/15/14659327 sigir program2:32 pm 1/15/14435451 sigir2014.org2:42 pm 1/15/14435451 information retrieval4:56 pm 1/15/14142039 sigir 20145:02 pm 1/15/14312055 xxx clubs on gold coast10:14 pm 1/15/13142039 sex videos1:49 am 1/16/13142039

20 QueryTimeUser sigir 201410:41 am 1/15/14142039 goldcoast sofitel10:44 am 1/15/14142039 teen sex10:56 am 1/15/14142039 sigir 201411:21 am 1/15/14659327 ool transportation11:59 am 1/15/14318222 restaurants brisbane12:01 pm 1/15/14318222 surf lessons12:17 pm 1/15/14318222 james allen12:18 pm 1/15/14142039 daytrips from brisbane1:30 pm 1/15/14554320 sex with animals1:30 pm 1/15/14659327 sigir program2:32 pm 1/15/14435451 sigir2014.org2:42 pm 1/15/14435451 Information retrieval4:56 pm 1/15/14142039 sigir 20145:02 pm 1/15/14312055 xxx clubs on gold coast10:14 pm 1/15/14142039 sex videos1:49 am 1/16/14142039 cheap digital camera12:17 pm 1/15/14554320 cheap digital camera12:18 pm 1/15/14554320 cheap digital camera12:19 pm 1/15/14554320 社会科学 11:59 am 11/3/23 12:01 pm 11/3/23 Porn Language Spam System errors

21 QueryTimeUser sigir 201410:41 am 1/15/14142039 goldcoast sofitel10:44 am 1/15/14142039 learning to rank10:56 am 1/15/14142039 sigir 201411:21 am 1/15/14659327 ool transportation11:59 am 1/15/14318222 restaurants brisbane12:01 pm 1/15/14318222 surf lessons12:17 pm 1/15/14318222 james allen12:18 pm 1/15/14142039 daytrips from brisbane1:30 pm 1/15/14554320 sigir 20141:30 pm 1/15/14659327 sigir program2:32 pm 1/15/14435451 sigir2014.org2:42 pm 1/15/14435451 information retrieval4:56 pm 1/15/14142039 sigir 20145:02 pm 1/15/14312055 kangaroos10:14 pm 1/15/14142039 machine learning1:49 am 1/16/14142039

22 QueryTimeUser sigir 201410:41 am 1/15/14142039 goldcoast sofitel10:44 am 1/15/14142039 learning to rank10:56 am 1/15/14142039 sigir 201411:21 am 1/15/14659327 ool transportation11:59 am 1/15/14318222 restaurants brisbane12:01 pm 1/15/14318222 surf lessons12:17 pm 1/15/14318222 james allen12:18 pm 1/15/14142039 daytrips from brisbane1:30 pm 1/15/14554320 sigir 20141:30 pm 1/15/14659327 sigir program2:32 pm 1/15/14435451 sigir2014.org2:42 pm 1/15/14435451 information retrieval4:56 pm 1/15/14142039 sigir 20145:02 pm 1/15/14312055 kangaroos10:14 pm 1/15/14142039 machine learning1:49 am 1/16/14142039 Query typology

23 QueryTimeUser sigir 201410:41 am 1/15/14142039 goldcoast sofitel10:44 am 1/15/14142039 learning to rank10:56 am 1/15/14142039 sigir 201411:21 am 1/15/14659327 ool transportation11:59 am 1/15/14318222 restaurants brisbane12:01 pm 1/15/14318222 surf lessons12:17 pm 1/15/14318222 james allen12:18 pm 1/15/14142039 daytrips from brisbane1:30 pm 1/15/14554320 sigir 20141:30 pm 1/15/14659327 sigir program2:32 pm 1/15/14435451 sigir2014.org2:42 pm 1/15/14435451 information retrieval4:56 pm 1/15/14142039 sigir 20145:02 pm 1/15/14312055 kangaroos10:14 pm 1/15/14142039 machine learning1:49 am 1/16/14142039 Query typology Query behavior

24 QueryTimeUser sigir 201410:41 am 1/15/14142039 goldcoast sofitel10:44 am 1/15/14142039 learning to rank10:56 am 1/15/14142039 sigir 201411:21 am 1/15/14659327 ool transportation11:59 am 1/15/14318222 restaurants brisbane12:01 pm 1/15/14318222 surf lessons12:17 pm 1/15/14318222 james allen12:18 pm 1/15/14142039 daytrips from brisbane1:30 pm 1/15/14554320 sigir 20141:30 pm 1/15/14659327 sigir program2:32 pm 1/15/14435451 sigir2014.org2:42 pm 1/15/14435451 information retrieval4:56 pm 1/15/14142039 sigir 20145:02 pm 1/15/14312055 kangaroos10:14 pm 1/15/14142039 machine learning1:49 am 1/16/14142039 Query typology Query behavior Long term trends Uses of Analysis Ranking – E.g., precision System design – E.g., caching User interface – E.g., history Test set development Complementary research

25 Surprises About Query Log Data  From early log analysis  Examples: Jansen et al. 2000, Broder 1998  Scale: Term common if it appeared 100 times!  Queries are not 7 or 8 words long  Advanced operators not used or “misused”  Nobody used relevance feedback  Lots of people search for sex  Navigation behavior common  Prior experience was with library search

26 Surprises About Microblog Search?

27 Ordered by time Ordered by time Ordered by relevance 8 new tweets Surprises About Microblog Search?

28 Ordered by time Ordered by time Ordered by relevance 8 new tweets Surprises About Microblog Search? Time important People important Specialized syntax Queries common Repeated a lot Change very little Often navigational Time and people less important No syntax use Queries longer Queries develop

29 Overview  Observational log analysis  What we can learn  Understand and predict user behavior  Collecting log data  Cleaning log data  Analyzing log data  Field observations

30 Collecting Log Data Observational Approaches to Information Retrieval

31 How to Get Logs for Analysis  Use existing logged data  Explore sources in your community (e.g., proxy logs)  Work with a company (e.g., FTE, intern, visiting researcher)  Generate your own logs  Focuses on questions of unique interest to you  Examples: UFindIt, Wikispeedia  Construct community resources  Shared software and tools  Client side logger (e.g., VIBE logger)  Shared data sets  Shared platform  Lemur Community Query Log Project

32 Web Service Logs Governmentcontractor Recruiting Academic field  Example sources  Search engine  Commercial site  Types of information  Queries, clicks, edits  Results, ads, products  Example analysis  Click entropy  Teevan, Dumais & Liebling. To Personalize or Not to Personalize: Modeling Queries with Variation in User Intent. SIGIR 2008

33 Controlled Web Service Logs  Example sources  Mechanical Turk  Games with a purpose  Types of information  Logged behavior  Active feedback  Example analysis  Search success  Ageev, Guo, Lagun & Agichtein. Find It If You Can: A Game for Modeling … Web Search Success Using Interaction Data. SIGIR 2011

34 Public Web Service Content  Example sources  Social network sites  Wiki change logs  Types of information  Public content  Dependent on service  Example analysis  Twitter topic models  Ramage, Dumais & Liebling. Characterizing microblogging using latent topic models. ICWSM 2010 j http://twahpic.cloudapp.net

35 Web Browser Logs  Example sources  Proxy  Logging tool  Types of information  URL visits, paths followed  Content shown, settings  Example analysis  DiffIE  Teevan, Dumais and Liebling. A Longitudinal Study of How Highlighting Web Content Change Affects.. Interactions. CHI 2010

36 Web Browser Logs  Example sources  Proxy  Logging tool  Types of information  URL visits, paths followed  Content shown, settings  Example analysis  Revisitation  Adar, Teevan and Dumais. Large Scale Analysis of Web Revisitation Patterns. CHI 2008

37 Rich Client-Side Logs  Example sources  Client application  Operating system  Types of information  Web client interactions  Other interactions – rich!  Example analysis  Stuff I’ve Seen  Dumais et al. Stuff I've Seen: A system for personal information retrieval and re-use. SIGIR 2003

38 dumaisbeijingsigir 2014vancouver A Simple Example  Logging search Queries and Clicked Results Web Service “SERP” chi 2014

39 A Simple Example  Logging Queries  Basic data:  Which time? time Client.send, time Server.receive, time Server.send, time Client.receive  Additional contextual data:  Where did the query come from?  What results were returned?  What algorithm or presentation was used?  Other metadata about the state of the system

40 A Simple Example  Logging Clicked Results (on the SERP)  How can a Web service know which SERP links are clicked?  Proxy re-direct  Script (e.g., JavaScript)  Dom and cross-browser challenges, but can instrument more than link clicks  No download required; but adds complexity and latency, and may influence user interaction  What happened after the result was clicked?  What happens beyond the SERP is difficult to capture  Browser actions (back, open in new tab, etc.) are difficult to capture  To better interpret user behavior, need richer client instrumentation http://www.chi2014.org vs. http://redir.service.com/?q=chi2014&url=http://www.chi2014.org/&pos= 3&log=DiFVYj1tRQZtv6e1FF7kltj02Z30eatB2jr8tJUFR function changeImage(){ document.imgC.src="thank_you..gif “; } function backImage(){ document.imgC.src=“image.gif"; }

41 A (Not-So-) Simple Example  Logging: Queries, Clicked Results, and Beyond

42 What to Log  Log as much as possible  Time keyed events, e.g.:  Ideal log allows user experience to be fully reconstructed  But … make reasonable choices  Richly instrumented client experiments can provide guidance  Consider the amount of data, storage required  Challenges with scale  Storage requirements  1k bytes/record x 10 records/query x 100 mil queries/day = 1000 Gb/day  Network bandwidth  Client to server; Data center to data center

43 What to Do with the Data  Keep as much raw data as possible  And allowable  Must consider Terms of Service, IRB  Post-process data to put into a usable form  Integrate across servers to organize the data  By time  By userID  Normalize time, URLs, etc.  Rich data cleaning

44 Practical Issues: Time  Time  Client time is closer to the user, but can be wrong or reset  Server time includes network latencies, but controllable  In both cases, need to synchronize time across multiple machines  Data integration  Ensure that joins of data are all using the same basis (e.g., UTC vs. local time)  Accurate timing data is critical for understanding the sequence of user activities, daily temporal patterns, etc.

45 Practical Issues: Users  Http cookies, IP address, temporary ID  Provides broad coverage and easy to use, but …  Multiple people use same machine  Same person uses multiple machines (and browsers)  How many cookies did you use today?  Lots of churn in these IDs  Jupiter Res (39% delete cookies monthly); Comscore (2.5x inflation)  Login or download client code (e.g., browser plug-in)  Better correspondence to people, but …  Requires sign-in or download  Results in a smaller and biased sample of people or data (who remember to login, decided to download, etc.)  Either way, loss of data

46 Using the Data Responsibly  What data is collected and how it can be used?  User agreements (terms of service)  Emerging industry standards and best practices  Trade-offs  More data:  More intrusive and potential privacy concerns, but also more useful for understanding interaction and improving systems  Less data:  Less intrusive, but less useful  Risk, benefit, and trust

47  August 4, 2006: Logs released to academic community  3 months, 650 thousand users, 20 million queries  Logs contain anonymized User IDs  August 7, 2006: AOL pulled the files, but already mirrored  August 9, 2006: New York Times identified Thelma Arnold  “A Face Is Exposed for AOL Searcher No. 4417749”  Queries for businesses, services in Lilburn, GA (pop. 11k)  Queries for Jarrett Arnold (and others of the Arnold clan)  NYT contacted all 14 people in Lilburn with Arnold surname  When contacted, Thelma Arnold acknowledged her queries  August 21, 2006: 2 AOL employees fired, CTO resigned  September, 2006: Class action lawsuit filed against AOL AnonIDQueryQueryTimeItemRankClickURL ----------------------------------------------------------- 1234567uw cse 2006-04-04 18:18:181http://www.cs.washington.edu/ 1234567uw admissions process2006-04-04 18:18:183http://admit.washington.edu/admission 1234567computer science hci2006-04-24 09:19:32 1234567 computer science hci2006-04-24 09:20:042http://www.hcii.cmu.edu 1234567 seattle restaurants2006-04-24 09:25:502http://seattletimes.nwsource.com/rests 1234567 perlman montreal2006-04-24 10:15:144http://oldwww.acm.org/perlman/guide.html 1234567uw admissions notification2006-05-20 13:13:13 … Example: AOL Search Dataset

48  Other well known AOL users  User 711391 i love alaska  http://www.minimovies.org/documentaires/view/ilovealaska  User 17556639 how to kill your wife  User 927  Anonymous IDs do not make logs anonymous  Contain directly identifiable information  Names, phone numbers, credit cards, social security numbers  Contain indirectly identifiable information  Example: Thelma’s queries  Birthdate, gender, zip code identifies 87% of Americans

49 Example: Netflix Challenge  October 2, 2006: Netflix announces contest  Predict people’s ratings for a $1 million dollar prize  100 million ratings, 480k users, 17k movies  Very careful with anonymity post-AOL  May 18, 2008: Data de-anonymized  Paper published by Narayanan & Shmatikov  Uses background knowledge from IMDB  Robust to perturbations in data  December 17, 2009: Doe v. Netflix  March 12, 2010: Netflix cancels second competition Ratings 1: [Movie 1 of 17770] 12, 3, 2006-04-18 [CustomerID, Rating, Date] 1234, 5, 2003-07-08 [CustomerID, Rating, Date] 2468, 1, 2005-11-12 [CustomerID, Rating, Date] … Movie Titles … 10120, 1982, “Bladerunner” 17690, 2007, “The Queen” … All customer identifying information has been removed; all that remains are ratings and dates. This follows our privacy policy... Even if, for example, you knew all your own ratings and their dates you probably couldn’t identify them reliably in the data because only a small sample was included (less than one tenth of our complete dataset) and that data was subject to perturbation.

50 Using the Data Responsibly  Control access to the data  Internally: Access control; data retention policy  Externally: Risky (e.g., AOL, Netflix, Enron, Facebook public)  Protect user privacy  Directly identifiable information  Social security, credit card, driver’s license numbers  Indirectly identifiable information  Names, locations, phone numbers … you’re so vain (e.g., AOL)  Putting together multiple sources indirectly (e.g., Netflix, hospital records)  Linking public and private data  k-anonymity; Differential privacy; etc.  Transparency and user control  Publicly available privacy policy  Give users control to delete, opt-out, etc.

51 Overview  Observational log analysis  What we can learn  Understand and predict user behavior  Collecting log data  Not as simple as it seems  Cleaning log data – Filip!  Analyzing log data  Field observations

52 [Filip on data cleaning] 52

53 Observational Approaches to Information Retrieval SIGIR 2014 Tutorial: Choices and Constraints (Part II) Diane Kelly, Filip Radlinski, Jaime Teevan

54 Overview  Observational log analysis  What we can learn  Understand and predict user behavior  Collecting log data  Not as simple as it seems  Cleaning log data  Significant portion of log analysis about cleaning  Analyzing log data  Field observations

55 Analyzing Log Data Observational Approaches to Information Retrieval

56 Develop Metrics to Capture Behavior [Joachims 2002] Sessions 2.20 queries long [Silverstein et al. 1999] [Lau and Horvitz, 1999] Navigational, Informational, Transactional [Broder 2002] 2.35 terms [Jansen et al. 1998] Queries appear 3.97 times [Silverstein et al. 1999]  Summary measures  Query frequency  Query length  Analysis of query intent  Query types and topics  Temporal features  Session length  Common re-formulations  Click behavior  Relevant results for query  Queries that lead to clicks

57 Develop Metrics to Capture Behavior Lee, Teevan, de la Chica. Characterizing multi-click search behavior. SIGIR 2014

58 Partitioning the Data  Language  Location  Time  User activity  Individual  Entry point  Device  System variant Baeza-Yates, Dupret, Velasco. A study of mobile search queries in Japan. WWW 2007

59 Partition by Time  Periodicities  Spikes  Real-time data  New behavior  Immediate feedback  Individual  Within session  Across sessions Beitzel, et al. Hourly analysis of a.. topically categorized web query log. SIGIR 2004

60 Partition by User  Temporary ID (e.g., cookie, IP address)  High coverage but high churn  Does not necessarily map directly to users  User account  Only a subset of users Teevan, Adar, Jones, Potts. Information re-retrieval: Repeat queries … SIGIR 2007

61 Partition by System Variant  Also known as controlled experiments  Some people see one variant, others another  Example: What color for search result links?  Bing tested 40 colors  Identified #0044CC  Value: $80 million

62 Considerations When Partitioning  Choose comparison groups carefully  From the same time period  With comparable users, tasks, etc.  Log a lot because it can be hard to recreate state  Which partition did a particular behavior fall into?  Confirm partitions with metrics that should be the same White, Dumais, Teevan. Characterizing the influence of domain expertise... WSDM 2009

63 Interpreting Significant Metrics  Often, everything is significant Adar, Teevan, Dumais. Large scale analysis of web revisitation patterns. CHI 2008

64 Interpreting Significant Metrics  Everything is significant, but not always meaningful  “All differences significant except when noted.”  Choose the metrics you care about first  Look for converging evidence  Look at the data  Beware: Typically very high variance  Large variance by user, task, noise  Calculate empirically

65 Confidence Intervals  Confidence interval (C.I.):  Interval around the treatment mean that contains the true value of the mean x% (typically 95%) of the time  Gives useful information about the size of the effect and its practical significance  C.I.s that do not contain the control mean are statistically significant (statistically different from the control)  This is an independent test for each metric  Thus you will get 1 in 20 results (for 95% C.I.s) that are spurious  Challenge: You don't know which ones are spurious

66 Confidence Intervals Lee, Teevan, de la Chica. Characterizing multi-click search behavior. SIGIR 2014 Radlinski, Kurup, Joachims. How does clickthrough data reflect retrieval quality? CIKM 2008.

67 When Significance Is Wrong  Sometimes there is spurious significance  Confidence interval only tells you there is a 95% chance that this difference is real; not 100%  If only a few things significant, chance a likely explanation  Sometimes you will miss significance  Because the true difference is tiny/zero or because you don’t have enough power  If you did your sizing right, you have enough power to see all the differences of practical significance  Sometimes reason for change is unexpected  Look at many metrics to get a big picture Chilton, Teevan. Addressing Info. Needs Directly in the Search Result Page. WWW 2011

68 Be Thoughtful When Combining Metrics  1995 and 1996 performance != Combined performance  Simpsons Paradox  Changes in mix (denominators) make combined metrics (ratios) inconsistent with yearly metrics Batting Average 19951996Combined HitsAt BatHitsAt BatHitsAt Bat Derek Jeter 1248183582195630.250.314.310 David Justice 10441145140149551.253.321.270

69 Detailed Analysis  Big Picture  Not all effects will point the same direction  Take a closer look at the items going in the “wrong” direction  Can you interpret them?  E.g., people are doing fewer next-pages because they are finding their answer on the first page  Could they be artifactual?  What if they are real?  What should be the impact on your conclusions? on your decision?  Significance and impact are not the same thing  Looking at % change vs. absolute change helps  Effect size depends on what you want to do with the data

70 Beware of Tyranny of the Data  Can provide insight into behavior  Example: What is search for, how needs are expressed  Can be used to test hypotheses  Example: Compare ranking variants or link color  Can only reveal what can be observed  Cannot tell you what you cannot observe  Example: Nobody uses Twitter to re-find

71  People’s intent  People’s success  People’s experience  People’s attention  People’s beliefs  Behavior can mean many things  81% of search sequences ambiguous [Viermetz et al. 2006] 7:16 – Try new engine What Logs Cannot Tell Us 7:16 – Read Result 1 7:20 – Read Result 3 7:27 –Save links locally 7:12 – Query 7:14 – Click Result 1 7:15 – Click Result 3

72 HCI Example: Click Entropy  Question: How ambiguous is a query?  Approach: Look at variation in clicks  Measure: Click entropy  Low if no variation  human computer …  High if lots of variation  hci Companies Wikipedia disambiguation HCI Teevan, Dumais, Liebling. To personalize or not to personalize... SIGIR 2008

73 Which Has Less Variation in Clicks?  www.usajobs.gov v. federal government jobs  find phone number v. msn live search  singapore pools v. singaporepools.com  tiffany v. tiffany’s  nytimes v. connecticut newspapers  campbells soup recipes v. vegetable soup recipe  soccer rules v. hockey equipment ? ? ? Results change Result quality varies Tasks impacts # of clicks Clicks/user = 1.1 Clicks/user = 2.1 Click position = 2.6 Click position = 1.6 Result entropy = 5.7 Result entropy = 10.7

74 Supplementing Log Data  Enhance log data  Collect associated information  Example: For browser logs, crawl visited webpages  Instrumented panels  Converging methods  Usability studies  Eye tracking  Surveys  Field studies  Diary studies

75  Large-scale log analysis of re-finding  Do people know they are re-finding?  Do they mean to re-find the result they do?  Why are they returning to the result?  Small-scale critical incident user study  Browser plug-in that logs queries and clicks  Pop up survey on repeat clicks and 1/8 new clicks  Insight into intent + Rich, real-world picture  Re-finding often targeted towards a particular URL  Not targeted when query changes or in same session Example: Re-Finding Intent Tyler, Teevan. Large scale query log analysis of re-finding. WSDM 2010

76 Example: Curious Browser  Browser plug-in to examine relationship between implicit and explicit behavior  Capture many implicit actions (e.g., click, click position, dwell time, scroll)  Probe for explicit user judgments of relevance of a page to the query  Deployed to ~4k people in US and Japan  Learned models to predict explicit judgments from implicit indicators  45% accuracy w/ just click; 75% accuracy w/ click + dwell + session  Used to identify important features; then apply model in open loop setting Fox, et al. Evaluating implicit measures to improve the search experience. TOIS 2005

77 Overview  Observational log analysis  What we can learn  Partition logs to observe behavior  Collecting log data  Not as simple as it seems  Cleaning log data  Clean and sanity check  Analyzing log data  Big picture more important than individual metrics  Field observations – Diane!

78 [Diane on field observations] 78


Download ppt "Observational Approaches to Information Retrieval SIGIR 2014 Tutorial: Choices and Constraints (Part II) Diane Kelly, Filip Radlinski, Jaime Teevan Slides."

Similar presentations


Ads by Google