Retroactive Answering of Search Queries Beverly Yang Glen Jeh Google.

Slides:



Advertisements
Similar presentations
Take a tour of De Gruyter Online
Advertisements

Jack Jedwab Association for Canadian Studies September 27 th, 2008 Canadian Post Olympic Survey.
1 IDX. 2 What you will learn: What IDX is Why its important How to use it Tips and tricks Introduction Q & A.
Copyright © 2003 Pearson Education, Inc.
Chapter 6 Flowcharting.
Beliefs & Biases in Web Search
Fatma Y. ELDRESI Fatma Y. ELDRESI ( MPhil ) Systems Analysis / Programming Specialist, AGOCO Part time lecturer in University of Garyounis,
Take another look Alison Hayman Search Solutions Unit Dissemination Divison February 2011 Statistics Canada site search.
1 Take a tour of De Gruyter Online Join the rally and learn how to navigate through our website.
OvidSP Flexible. Innovative. Precise. Introducing OvidSP Resources.
Effective Change Detection Using Sampling Junghoo John Cho Alexandros Ntoulas UCLA.
LIS618 lecture 2 Thomas Krichel Structure Theory: information retrieval performance Practice: more advanced dialog.
We need a common denominator to add these fractions.
Exploring Traversal Strategy for Web Forum Crawling Yida Wang, Jiang-Ming Yang, Wei Lai, Rui Cai, Lei Zhang and Wei-Ying Ma Chinese Academy of Sciences.
Jeopardy Q 1 Q 6 Q 11 Q 16 Q 21 Q 2 Q 7 Q 12 Q 17 Q 22 Q 3 Q 8 Q 13
Serving the interests of e-resource users in Research, Higher and Further Education Libraries Librarians and research evaluation support overview.
Internet Search Engine freshness by Web Server help Presented by: Barilari Alessandro.
LIBRARY WEBSITE, CATALOG, DATABASES AND FREE WEB RESOURCES.
PhishZoo: Detecting Phishing Websites By Looking at Them
Why Blog? To market or promote yourself, service or product.
A Fractional Order (Proportional and Derivative) Motion Controller Design for A Class of Second-order Systems Center for Self-Organizing Intelligent.
Google News Personalization: Scalable Online Collaborative Filtering
Filtering Semi-Structured Documents Based on Faceted Feedback Lanbo Zhang, Yi Zhang, Qianli Xing Information Retrieval and Knowledge Management (IRKM)
Evaluating the Robustness of Learning from Implicit Feedback Filip Radlinski Thorsten Joachims Presentation by Dinesh Bhirud
1 SESSION 5- RECORDING AND REPORTING IN GRADES R-12 Computer Applications Technology Information Technology.
The SCPS Professional Growth System
The basics for simulations
Learning to Suggest: A Machine Learning Framework for Ranking Query Suggestions Date: 2013/02/18 Author: Umut Ozertem, Olivier Chapelle, Pinar Donmez,
1 Challenge the future Subtitless On Lightweight Design of Submarine Pressure Hulls.
Well-Being Icon Refer to Slide 2 for instructions on how to view the full-screen slideshow.Slide 2.
TCCI Barometer March “Establishing a reliable tool for monitoring the financial, business and social activity in the Prefecture of Thessaloniki”
Dynamic Access Control the file server, reimagined Presented by Mark on twitter 1 contents copyright 2013 Mark Minasi.
1 Evaluations in information retrieval. 2 Evaluations in information retrieval: summary The following gives an overview of approaches that are applied.
1 3GPP TSGs SA Meeting #49, San Antonio, Texas, USA, September 2010 SP © 3GPP Organizational Partners Satisfaction survey results In.
Promoting Regulatory Excellence Self Assessment & Physiotherapy: the Ontario Model Jan Robinson, Registrar & CEO, College of Physiotherapists of Ontario.
Information Retrieval (IR) on the Internet. Contents  Definition of IR  Performance Indicators of IR systems  Basics of an IR system  Some IR Techniques.
CMPT 275 Software Engineering
TCCI Barometer September “Establishing a reliable tool for monitoring the financial, business and social activity in the Prefecture of Thessaloniki”
Advanced Web Metrics with Google Analytics By: Carley Brown.
Slide 01 (of 22)Title 26/04/2010 Version 1.0 GUIDE to ‘SIMPLE’ Mouse click to continue AN OVERVIEW OF BT’s CONVEYACE INVOICE RECONCILIATION ASSISTANCE.
]po[ Docu Wiki.  ]project-opem[ 2008, Rollout Methodology / Frank Bergmann / 2 Types of Readers  Beginners – These users have just started using ]po[.
2011 WINNISQUAM COMMUNITY SURVEY YOUTH RISK BEHAVIOR GRADES 9-12 STUDENTS=1021.
Before Between After.
Benjamin Banneker Charter Academy of Technology Making AYP Benjamin Banneker Charter Academy of Technology Making AYP.
2011 FRANKLIN COMMUNITY SURVEY YOUTH RISK BEHAVIOR GRADES 9-12 STUDENTS=332.
Requirements Analysis 1. 1 Introduction b501.ppt © Copyright De Montfort University 2000 All Rights Reserved INFO2005 Requirements Analysis Introduction.
Subtraction: Adding UP
Music Recommendation by Unified Hypergraph: Music Recommendation by Unified Hypergraph: Combining Social Media Information and Music Content Jiajun Bu,
1 Teaching the Web in Under an Hour Mary Ellen Bates Bates Information Services
Week 1.
Middle School Lesson 2 Activity 3 – The Guessing Game
DIKLA GRUTMAN 2014 Databases- presentation and training.
Uktradeinfo Web Survey Results Published: April 2014.
INFORMATION SOLUTIONS Citation Analysis Reports. Copyright 2005 Thomson Scientific 2 INFORMATION SOLUTIONS Provide highly customized datasets based on.
Patient Survey Results 2013 Nicki Mott. Patient Survey 2013 Patient Survey conducted by IPOS Mori by posting questionnaires to random patients in the.
Contract Audit Follow-Up (CAFU) 3.5 Pre-Defined & Ad hoc Reports November 2009 ITCSO Training Academy.
1 Distributed Agents for User-Friendly Access of Digital Libraries DAFFODIL Effective Support for Using Digital Libraries Norbert Fuhr University of Duisburg-Essen,
RollCaller: User-Friendly Indoor Navigation System Using Human-Item Spatial Relation Yi Guo, Lei Yang, Bowen Li, Tianci Liu, Yunhao Liu Hong Kong University.
Psychological Advertising: Exploring User Psychology for Click Prediction in Sponsored Search Date: 2014/03/25 Author: Taifeng Wang, Jiang Bian, Shusen.
Optimizing search engines using clickthrough data
WWW 2014 Seoul, April 8 th SNOW 2014 Data Challenge Two-level message clustering for topic detection in Twitter Georgios Petkos, Symeon Papadopoulos, Yiannis.
J. Chen, O. R. Zaiane and R. Goebel An Unsupervised Approach to Cluster Web Search Results based on Word Sense Communities.
Improving Web Search Ranking by Incorporating User Behavior Information Eugene Agichtein Eric Brill Susan Dumais Microsoft Research.
When Experts Agree: Using Non-Affiliated Experts To Rank Popular Topics Meital Aizen.
Retroactive Answering of Search Queries Beverly Yang Glen Jeh.
“In the beginning -- before Google -- a darkness was upon the land.” Joel Achenbach Washington Post.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
INFORMATION RETRIEVAL MEASUREMENT OF RELEVANCE EFFECTIVENESS 1Adrienn Skrop.
Jaime Teevan MIT, CSAIL The Re:Search Engine. “Pick a card, any card.”
Retroactive Answering of Search Queries Beverly Yang Glen Jeh Google, Inc. Presented.
Presentation transcript:

Retroactive Answering of Search Queries Beverly Yang Glen Jeh Google

Personalization Provide more relevant services to specific user Based on Search History Usually operates at a high level e.g., Re-order search results based on a users general preferences Classic example: User likes cars Query: jaguar Why not focus on known, specific needs? User likes cars User is interested in the 2006 Honda Civic

The QSR system QSR = Query-Specific (Web) Recommendations Alerts user when interesting new results to selected previous queries have appeared Example Query: britney spears concert san francisco No good results at time of query (Britney not on tour) One month later, new results (Britney is coming to town!) User is automatically notified

Query treated as standing query New results are web page recommendations

Challenges How do we identify queries representing standing interests? Explicit – Web Alerts. But no one does this Want to automatically identify How do we identify interesting new results? Web alerts: change in top 10. But thats not good enough Focus: addressing these two challenges

Outline Introduction Basic QSR Architecture Identifying Standing Interests Determining Interesting Results User Study Setup Results Heuristic

Architecture Search Engine History Database Actions QSR Engine (1) Identify Interests (2) Identify New Results Actions Queries Recommendations Limit: M queries

Related Work Identifying User Goal [Rose & Levinson 2004], [Lee, Liu & Cho 2005] At a higher, more general level Identifying Satisfaction [Fox, et. al. 2005] One component of identifying standing interest Specific model, holistic rather than considering strength and characteristics of each signal Recommendation Systems Too many to list!

Outline Introduction Basic QSR Architecture Identifying Standing Interests Determining Interesting Results User Study Setup Results

Definition A user has a standing interest in a query if she would be interested in seeing new interesting results Factors to consider: Prior fulfillment/Satisfaction Query interest level Duration of need or interest

Example QUERY (8s) -- html encode java RESULTCLICK (91s) – 2. RESULTCLICK (247s) – 1. RESULTCLICK (12s) – 8. NEXTPAGE (5s) – start = 10 RESULTCLICK (1019s) – REFINEMENT (21s) – html encode java utility RESULTCLICK (32s) – 7. NEXTPAGE (8s) – start = 10 NEXTPAGE (30s) – start = 20

Example QUERY (8s) -- html encode java RESULTCLICK (91s) – 2. RESULTCLICK (247s) – 1. RESULTCLICK (12s) – 8. NEXTPAGE (5s) – start = 10 RESULTCLICK (1019s) – REFINEMENT (21s) – html encode java utility RESULTCLICK (32s) – 7. NEXTPAGE (8s) – start = 10 NEXTPAGE (30s) – start = 20

Signals Good ones: # terms # clicks, # refinements History match Repeated non-navigational Other: Session duration, number of long clicks, etc.

Outline Introduction Basic QSR Architecture Identifying Standing Interests Determining Interesting Results User Study Setup Results

Web Alerts Heuristic: new result in top 10 Query: beverly yang Alert 10/16/2005: Seen before through a web search Poor quality page Alert repeated due to ranking fluctuations

QSR Example RankURLPR scoreSeen 1www.rssreader.com3.93Yes 2blogspace.com/rss/readers3.19Yes 3www.feedreader.com3.23Yes 4www.google.com/reader2.74No 5www.bradsoft.com2.80Yes 6www.bloglines.com2.84Yes 7www.pluck.com2.63Yes 8sage.mozdev.org2.56Yes 9www.sharpreader.net2.61Yes Query: rss reader (not real)

Signals Good ones: History presence Rank (inverse!) Popularity and relevance (PR) scores Above dropoff PR scores of a few results are much higher than PR scores of the rest Content match Other: Days elapsed since query, sole changed

Outline Introduction Basic QSR Architecture Identifying Standing Interests Determining Interesting Results User Study Setup Results

Overview Human subjects: Google Search History users Purpose: Demonstrate promise of system effectiveness Verify intuitions behind heuristics Many disclaimers: Study conducted internally!!! 18 subjects!!! Only a fraction of queries in each subjects history!!! Need additional studies over broader populations to generalize results

Questionnaire QUERY (8s) -- html encode java RESULTCLICK (91s) – 2. RESULTCLICK (247s) – 1. RESULTCLICK (12s) – 8. NEXTPAGE (5s) – start = 10 RESULTCLICK (1019s) – REFINEMENT (21s) – html encode java utility RESULTCLICK (32s) – 7. NEXTPAGE (8s) – start = 10 NEXTPAGE (30s) – start = 20 1)Did you find a satisfactory answer for your query? Yes Somewhat No Cant Remember 2)How interested would you be in seeing a new high-quality result? Very Somewhat Vaguely Not 3)How long would this interest last for? Ongoing Month Week Now 4)How good would you rate the quality of this result? Excellent Good Fair Poor

Outline Introduction Basic QSR Architecture Identifying Standing Interests Determining Interesting Results User Study Setup Results

Questions Is there a need for automatic detection of standing interests? Which signals are useful for indicating standing interest in a query session? Which signals are useful for indicating quality of recommendations?

Is there a need? How many Web alerts have you ever registered? Of the queries marked very or somewhat interesting (154 total), how many have you registered? 0: 73% 1: 20% 2: 7% >2: 0% 0: 100%

Effectiveness of Signals Standing interests # clicks (> 8) # refinements (> 3) History match Also: repeated non-navigational, # terms (> 2) Quality Results PR score (high) Rank (low!!) Above Dropoff

Standing Interest

Prior Fulfillment

Interest Score Goal: capture the relative standing interest a user has in a query session iscore = a * log(# clicks + # refinements) + b * log(# repetitions) + c * (history match score) Select query sessions with iscore > t

Effectiveness of iscore Standing Interest: Sessions for which user is somewhat or very interested in seeing further results Select query sessions with iscore > t Vary t to get precision/recall tradeoff 90% precision, 11% recall 69% precision, 28% recall Compare: 28% precision by random selection Recall – percentage of standing interest sessions that appeared in the survey

Quality of Results Desired: marked in survey as good or excellent

Quality Score Goal: capture relative quality of recommendation Apply score after result has passed a number of boolean filters qscore = a * PR score + b * rank c * topic match 1 b * ---- rank

Effectiveness of qscore Recall: Percentage of URLs in the survey marked as good or excellent Select URLs with score > t

Conclusion Huge gap: Users standing interests/needs Existing technology to address them QSR: Retroactively answer search queries Automatic identification of standing interests and unfulfilled needs Identification of interesting new results Future work Broader studies Feedback loop

Thank you!

Selecting Sessions Users may have thousands of queries Must only show 30 Try to include a mix of positive and negative sessions Prevents us from gathering some stats Process Filter special-purpose queries (e.g., maps) Filter sessions with 1-2 actions Rank sessions by iscore Take top 15 sessions by score Take 15 randomly chosen sessions

Selecting Recommendations Tried to only show good recommendations Assumption: some will be bad Process Only consider sessions with history presence Only consider results in top 10 (Google) Must pass at least 2 boolean signals Select top 50% according to qscore

3 rd -Person study Not enough recommendations in 1 st - person study Asked subjects to evaluate recommendations made for other users sessions