Mining the Search Trails of Surfing Crowds: Identifying Relevant Websites from User Activity Data Misha Bilenko and Ryen White presented by Matt Richardson.

Slides:



Advertisements
Similar presentations
Predicting User Interests from Contextual Information
Advertisements

Evaluating the Robustness of Learning from Implicit Feedback Filip Radlinski Thorsten Joachims Presentation by Dinesh Bhirud
Struggling or Exploring? Disambiguating Long Search Sessions
Improvements and extras Paul Thomas CSIRO. Overview of the lectures 1.Introduction to information retrieval (IR) 2.Ranked retrieval 3.Probabilistic retrieval.
1 Evaluation Rong Jin. 2 Evaluation  Evaluation is key to building effective and efficient search engines usually carried out in controlled experiments.
Modelling Relevance and User Behaviour in Sponsored Search using Click-Data Adarsh Prasad, IIT Delhi Advisors: Dinesh Govindaraj SVN Vishwanathan* Group:
Optimizing search engines using clickthrough data
1.Accuracy of Agree/Disagree relation classification. 2.Accuracy of user opinion prediction. 1.Task extraction performance on Bing web search log with.
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
Catching the Drift: Learning Broad Matches from Clickthrough Data Sonal Gupta, Mikhail Bilenko, Matthew Richardson University of Texas at Austin, Microsoft.
1 Learning User Interaction Models for Predicting Web Search Result Preferences Eugene Agichtein Eric Brill Susan Dumais Robert Ragno Microsoft Research.
Click Evidence Signals and Tasks Vishwa Vinay Microsoft Research, Cambridge.
Personalizing Search via Automated Analysis of Interests and Activities Jaime Teevan Susan T.Dumains Eric Horvitz MIT,CSAILMicrosoft Researcher Microsoft.
Learning to Rank: New Techniques and Applications Martin Szummer Microsoft Research Cambridge, UK.
Presented by Li-Tal Mashiach Learning to Rank: A Machine Learning Approach to Static Ranking Algorithms for Large Data Sets Student Symposium.
Ryen W. White, Microsoft Research Jeff Huang, University of Washington.
1 LM Approaches to Filtering Richard Schwartz, BBN LM/IR ARDA 2002 September 11-12, 2002 UMASS.
An investigation of query expansion terms Gheorghe Muresan Rutgers University, School of Communication, Information and Library Science 4 Huntington St.,
Scalable Text Mining with Sparse Generative Models
Information Retrieval
Overview of Search Engines
Cohort Modeling for Enhanced Personalized Search Jinyun YanWei ChuRyen White Rutgers University Microsoft BingMicrosoft Research.
Adapting Deep RankNet for Personalized Search
Modern Retrieval Evaluations Hongning Wang
Query Rewriting Using Monolingual Statistical Machine Translation Stefan Riezler Yi Liu Google 2010 Association for Computational Linguistics.
Reyyan Yeniterzi Weakly-Supervised Discovery of Named Entities Using Web Search Queries Marius Pasca Google CIKM 2007.
User Browsing Graph: Structure, Evolution and Application Yiqun Liu, Yijiang Jin, Min Zhang, Shaoping Ma, Liyun Ru State Key Lab of Intelligent Technology.
PageRank for Product Image Search Kevin Jing (Googlc IncGVU, College of Computing, Georgia Institute of Technology) Shumeet Baluja (Google Inc.) WWW 2008.
Understanding and Predicting Graded Search Satisfaction Tang Yuk Yu 1.
Improving Web Search Ranking by Incorporating User Behavior Information Eugene Agichtein Eric Brill Susan Dumais Microsoft Research.
Ramakrishnan Srikant Sugato Basu Ni Wang Daryl Pregibon 1.
Fan Guo 1, Chao Liu 2 and Yi-Min Wang 2 1 Carnegie Mellon University 2 Microsoft Research Feb 11, 2009.
Web Search. Structure of the Web n The Web is a complex network (graph) of nodes & links that has the appearance of a self-organizing structure  The.
Ryen W. White, Matthew Richardson, Mikhail Bilenko Microsoft Research Allison Heath Rice University.
Implicit User Feedback Hongning Wang Explicit relevance feedback 2 Updated query Feedback Judgments: d 1 + d 2 - d 3 + … d k -... Query User judgment.
Context-Sensitive Information Retrieval Using Implicit Feedback Xuehua Shen : department of Computer Science University of Illinois at Urbana-Champaign.
Ryen W. White, Dan Morris Microsoft Research, Redmond, USA {ryenw,
Question Answering over Implicitly Structured Web Content
Information Retrieval Effectiveness of Folksonomies on the World Wide Web P. Jason Morrison.
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
Personalizing Web Search using Long Term Browsing History Nicolaas Matthijs, Cambridge Filip Radlinski, Microsoft In Proceedings of WSDM
Personalization with user’s local data Personalizing Search via Automated Analysis of Interests and Activities 1 Sungjick Lee Department of Electrical.
Qi Guo Emory University Ryen White, Susan Dumais, Jue Wang, Blake Anderson Microsoft Presented by Tetsuya Sakai, Microsoft Research.
Adish Singla, Microsoft Bing Ryen W. White, Microsoft Research Jeff Huang, University of Washington.
Research © 2008 Yahoo! Generating Succinct Titles for Web URLs Kunal Punera joint work with Deepayan Chakrabarti and Ravi Kumar Yahoo! Research.
Implicit User Feedback Hongning Wang Explicit relevance feedback 2 Updated query Feedback Judgments: d 1 + d 2 - d 3 + … d k -... Query User judgment.
Date: 2012/11/29 Author: Chen Wang, Keping Bi, Yunhua Hu, Hang Li, Guihong Cao Source: WSDM’12 Advisor: Jia-ling, Koh Speaker: Shun-Chen, Cheng.
Modern Retrieval Evaluations Hongning Wang
DISTRIBUTED INFORMATION RETRIEVAL Lee Won Hee.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Date: 2013/9/25 Author: Mikhail Ageev, Dmitry Lagun, Eugene Agichtein Source: SIGIR’13 Advisor: Jia-ling Koh Speaker: Chen-Yu Huang Improving Search Result.
RapStar’s Solution to Data Mining Hackathon on Best Buy Mobile Site Kingsfield, Dragon.
Why Decision Engine Bing Demos Search Interaction model Data-driven Research Problems Q & A.
Predicting User Interests from Contextual Information R. W. White, P. Bailey, L. Chen Microsoft (SIGIR 2009) Presenter : Jae-won Lee.
Learning to Rank: From Pairwise Approach to Listwise Approach Authors: Zhe Cao, Tao Qin, Tie-Yan Liu, Ming-Feng Tsai, and Hang Li Presenter: Davidson Date:
Predicting Short-Term Interests Using Activity-Based Search Context CIKM’10 Advisor: Jia Ling, Koh Speaker: Yu Cheng, Hsieh.
Navigation Aided Retrieval Shashank Pandit & Christopher Olston Carnegie Mellon & Yahoo.
N-Gram Model Formulas Word sequences Chain rule of probability Bigram approximation N-gram approximation.
Usefulness of Quality Click- through Data for Training Craig Macdonald, ladh Ounis Department of Computing Science University of Glasgow, Scotland, UK.
Assessing the Scenic Route: Measuring the Value of Search Trails in Web Logs Ryen W. White1 Jeff Huang2 1Microsoft Research 1University of Washington.
Search Engine Architecture
Search User Behavior: Expanding The Web Search Frontier
Content-Aware Click Modeling
Eugene Agichtein Mathematics & Computer Science Emory University
Evidence from Behavior
CS246: Leveraging User Feedback
Date: 2012/11/15 Author: Jin Young Kim, Kevyn Collins-Thompson,
Query Type Classification for Web Document Retrieval
INF 141: Information Retrieval
Presentation transcript:

Mining the Search Trails of Surfing Crowds: Identifying Relevant Websites from User Activity Data Misha Bilenko and Ryen White presented by Matt Richardson Microsoft Research

Search = Modeling User Behavior Retrieval functions estimate relevance from behavior of several user groups: – Page authors create page contents TF-IDF/BM25, query-is-page-title, … – Page authors create links PageRank/HITS, query-matches-anchor text, … – Searchers submit queries and click on results Clickthrough, query reformulations Most user behavior occurs beyond search engines – Viewing results and browsing beyond them – What can we capture, and how can we use it?

Prior Work Clickthrough/implicit feedback methods – Learning ranking functions from clicks and query chains [Joachims ‘02, Xue et al. ‘04, Radlinski-Joachims ’05 ‘06 ‘07] – Combining clickthrough with traditional IR features [Richardson et al. ‘06, Agichtein et al. ‘06] Activity-based user models for personalization – [Shen et al. ‘05, Tan et al. ’06] Modeling browsing behavior – [Anderson et al. ‘01, Downey et al. ‘07, Pandit-Olston ’07]

Search Trails Trails start with a search engine query Continue until a terminating event – Another search – Visit to an unrelated site (social networks, webmail) – Timeout, browser homepage, browser closing

Trails vs. Click logs Trails capture dwell time Both attention share and pageview counts are accounted Trails represent user activity across many websites Browsing sequences surface “under-ranked” pages Click logs are less noisy Position bias is easy to control

Predicting Relevance from Trails Task: given a trails corpus D ={q i → (d i1,…,d ik )} predict relevant websites for a new query q Trails give us the good pages for each query… …can’t we just lookup the pages for new queries? – Not directly: 50+% of queries are unique – Page visits are also extremely sparse Solutions: – Query sparsity: term-based matching, language modeling – Pageview sparsity: smoothing (domain-level prediction)

Model 1: Heuristic Documents ≈ websites Contents ≈ queries preceding websites in trails Split queries into terms, compute frequencies – Terms include unigrams, bigrams, named entities Relevance is analogous to BM25 (TF-IDF) – Query-term frequency (QF) and inverse query frequency (IQF) terms incorporate corpus statistics and website popularity.

Model 2: Probabilistic IR via language modeling [Zhai-Lafferty, Lavrenko] Query-term distribution gives more mass to rare terms: Term-website weights combine dwell time and counts

Model 2: Probabilistic (cont.) Basic probabilistic model is noisy – Misspellings, synonyms, sparseness

Model 3: Random Walks Basic probabilistic model is noisy – Misspellings, synonyms, sparseness Solution: random walk extension

Evaluation Train: 140+ million search trails (toolbar data) Test: human-labeled relevance set, 33K queries q =[black diamond carabiners] URLRating

Evaluation (cont.) Metric: NDCG (N ormalized D iscounted C umulative G ain ) Preferable to MAP, Kendall’s Tau, Spearman’s, etc. – Sensitive to top-ranked results – Handles variable number of results/target items – Well correlated with user satisfaction [Bompada et al. ‘07]

Evaluation (cont.) Metric: NDCG (N ormalized D iscounted C umulative G ain ) idr(i)DCG perfect (i) 1d1d d2d d3d d4d d5d idr(i)DCG(i)NDCG(i) 1d1d d7d d4d d5d d2d Perfect rankingObtained ranking

Results I: Domain ranking (cont.) Predicting correct ranking of domains for queries

Results I: Domain ranking (cont.) Full trails vs. search result clicks vs. “destinations”

Results I: Domain ranking (cont.) Scoring based on dwell times vs. visitation counts

Results I: Domain ranking (cont.) What’s better than data? LOTS OF DATA!

Results II: Learning to Rank Add Rel(q, d i ) as a feature to RankNet [Burges et al. ‘05] – Thousands of other features capture various content-, link- and clickthrough-based evidence

Conclusions Post-search browsing behavior (search trails) can be mined to extract users’ implicit endorsement of relevant websites. Trail-based relevance prediction provides unique signal not captured by other (content, link, clickthrough) features. Using full trails outperforms using only search result clicks or search trail destinations. Probabilistic models incorporating random walks provide best accuracy by overcoming data sparsity and noise.

Model 3: Random Walks (cont.)

URLs vs. Websites Website ≈ domain – Sites: spaces.live.com, news.yahoo.co.uk – Not sites: www2.hp.com, cx09hz.myspace.com Scoring: URLRating URLRating bdel.comPerfect rei.comGood blackdiamondranch.comBad URL rankingWebsite ranking