Catching the Drift: Learning Broad Matches from Clickthrough Data Sonal Gupta, Mikhail Bilenko, Matthew Richardson University of Texas at Austin, Microsoft.

Slides:



Advertisements
Similar presentations
Context-Sensitive Query Auto-Completion AUTHORS:NAAMA KRAUS AND ZIV BAR-YOSSEF DATE OF PUBLICATION:NOVEMBER 2010 SPEAKER:RISHU GUPTA 1.
Advertisements

Online Ad Auctions By : Hal R. Varian Reviewed By : Sahil Gupta Instructor : Professor Mattmann TA: Kaijan Xu 14/11/2013.
Evaluating the Robustness of Learning from Implicit Feedback Filip Radlinski Thorsten Joachims Presentation by Dinesh Bhirud
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Improvements and extras Paul Thomas CSIRO. Overview of the lectures 1.Introduction to information retrieval (IR) 2.Ranked retrieval 3.Probabilistic retrieval.
Psychological Advertising: Exploring User Psychology for Click Prediction in Sponsored Search Date: 2014/03/25 Author: Taifeng Wang, Jiang Bian, Shusen.
1 Evaluation Rong Jin. 2 Evaluation  Evaluation is key to building effective and efficient search engines usually carried out in controlled experiments.
Modelling Relevance and User Behaviour in Sponsored Search using Click-Data Adarsh Prasad, IIT Delhi Advisors: Dinesh Govindaraj SVN Vishwanathan* Group:
Optimizing search engines using clickthrough data
Query Chains: Learning to Rank from Implicit Feedback Paper Authors: Filip Radlinski Thorsten Joachims Presented By: Steven Carr.
WSCD INTRODUCTION  Query suggestion has often been described as the process of making a user query resemble more closely the documents it is expected.
Jean-Eudes Ranvier 17/05/2015Planet Data - Madrid Trustworthiness assessment (on web pages) Task 3.3.
1 Learning User Interaction Models for Predicting Web Search Result Preferences Eugene Agichtein Eric Brill Susan Dumais Robert Ragno Microsoft Research.
Context-aware Query Suggestion by Mining Click-through and Session Data Authors: H. Cao et.al KDD 08 Presented by Shize Su 1.
Personalizing Search via Automated Analysis of Interests and Activities Jaime Teevan Susan T.Dumains Eric Horvitz MIT,CSAILMicrosoft Researcher Microsoft.
Time-dependent Similarity Measure of Queries Using Historical Click- through Data Qiankun Zhao*, Steven C. H. Hoi*, Tie-Yan Liu, et al. Presented by: Tie-Yan.
A machine learning approach to improve precision for navigational queries in a Web information retrieval system Reiner Kraft
Scott Wen-tau Yih Joint work with Kristina Toutanova, John Platt, Chris Meek Microsoft Research.
1 Today  Tools (Yves)  Efficient Web Browsing on Hand Held Devices (Shrenik)  Web Page Summarization using Click- through Data (Kathy)  On the Summarization.
A Search-based Method for Forecasting Ad Impression in Contextual Advertising Defense.
Personalized Ontologies for Web Search and Caching Susan Gauch Information and Telecommunications Technology Center Electrical Engineering and Computer.
Finding Advertising Keywords on Web Pages Scott Wen-tau YihJoshua Goodman Microsoft Research Vitor R. Carvalho Carnegie Mellon University.
Computational Advertising Duygu Gunaydin Lu Li Shuanglong Wei.
Advisor: Hsin-Hsi Chen Reporter: Chi-Hsin Yu Date:
Tag-based Social Interest Discovery
Personalization in Local Search Personalization of Content Ranking in the Context of Local Search Philip O’Brien, Xiao Luo, Tony Abou-Assaleh, Weizheng.
Query Rewriting Using Monolingual Statistical Machine Translation Stefan Riezler Yi Liu Google 2010 Association for Computational Linguistics.
Anindya Ghose Sha Yang Stern School of Business New York University An Empirical Analysis of Sponsored Search Performance in Search Engine Advertising.
 An important problem in sponsored search advertising is keyword generation, which bridges the gap between the keywords bidded by advertisers and queried.
1 Cross-Lingual Query Suggestion Using Query Logs of Different Languages SIGIR 07.
Improving Web Search Ranking by Incorporating User Behavior Information Eugene Agichtein Eric Brill Susan Dumais Microsoft Research.
Ramakrishnan Srikant Sugato Basu Ni Wang Daryl Pregibon 1.
Fan Guo 1, Chao Liu 2 and Yi-Min Wang 2 1 Carnegie Mellon University 2 Microsoft Research Feb 11, 2009.
Xiaoying Gao Computer Science Victoria University of Wellington Intelligent Agents COMP 423.
UOS 1 Ontology Based Personalized Search Zhang Tao The University of Seoul.
Online Advertising Greg Lackey. Advertising Life Cycle The Past Mass media Current Media fragmentation The Future Target market Audio/visual enhancements.
Giorgos Giannopoulos (IMIS/”Athena” R.C and NTU Athens, Greece) Theodore Dalamagas (IMIS/”Athena” R.C., Greece) Timos Sellis (IMIS/”Athena” R.C and NTU.
Implicit User Feedback Hongning Wang Explicit relevance feedback 2 Updated query Feedback Judgments: d 1 + d 2 - d 3 + … d k -... Query User judgment.
Personalized Search Xiao Liu
Date: 2012/3/5 Source: Marcus Fontouraet. al(CIKM’11) Advisor: Jia-ling, Koh Speaker: Jiun Jia, Chiou 1 Efficiently encoding term co-occurrences in inverted.
Context-Sensitive Information Retrieval Using Implicit Feedback Xuehua Shen : department of Computer Science University of Illinois at Urbana-Champaign.
CONVERSION ARCHITECTURE CONVERSION ARCHITECTURE Testing data Keyword expansion Historical data Conversion analysis Geographic data Keyword analysis Visual.
Contextual Ranking of Keywords Using Click Data Utku Irmak, Vadim von Brzeski, Reiner Kraft Yahoo! Inc ICDE 09’ Datamining session Summarized.
Deep Learning Powered In- Session Contextual Ranking using Clickthrough Data Xiujun Li 1, Chenlei Guo 2, Wei Chu 2, Ye-Yi Wang 2, Jude Shavlik 1 1 University.
IR Theory: Relevance Feedback. Relevance Feedback: Example  Initial Results Search Engine2.
 Examine two basic sources for implicit relevance feedback on the segment level for search personalization. Eye tracking Display time.
Modeling term relevancies in information retrieval using Graph Laplacian Kernels Shuguang Wang Joint work with Saeed Amizadeh and Milos Hauskrecht.
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
Personalizing Web Search using Long Term Browsing History Nicolaas Matthijs, Cambridge Filip Radlinski, Microsoft In Proceedings of WSDM
Personalization with user’s local data Personalizing Search via Automated Analysis of Interests and Activities 1 Sungjick Lee Department of Electrical.
Mining Dependency Relations for Query Expansion in Passage Retrieval Renxu Sun, Chai-Huat Ong, Tat-Seng Chua National University of Singapore SIGIR2006.
Post-Ranking query suggestion by diversifying search Chao Wang.
Context-Aware Query Classification Huanhuan Cao, Derek Hao Hu, Dou Shen, Daxin Jiang, Jian-Tao Sun, Enhong Chen, Qiang Yang Microsoft Research Asia SIGIR.
Generating Query Substitutions Alicia Wood. What is the problem to be solved?
Personalization Services in CADAL Zhang yin Zhuang Yuting Wu Jiangqin College of Computer Science, Zhejiang University November 19,2006.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Learning to Rank: From Pairwise Approach to Listwise Approach Authors: Zhe Cao, Tao Qin, Tie-Yan Liu, Ming-Feng Tsai, and Hang Li Presenter: Davidson Date:
1 Learning to Impress in Sponsored Search Xin Supervisors: Prof. King and Prof. Lyu.
CS791 - Technologies of Google Spring A Web­based Kernel Function for Measuring the Similarity of Short Text Snippets By Mehran Sahami, Timothy.
Ariel Fuxman, Panayiotis Tsaparas, Kannan Achan, Rakesh Agrawal (2008) - Akanksha Saxena 1.
Using Blog Properties to Improve Retrieval Gilad Mishne (ICWSM 2007)
Information Organization: Overview
An Empirical Study of Learning to Rank for Entity Search
Reach People when it matters with Location Extensions
FRM: Modeling Sponsored Search Log with Full Relational Model
Learning Term-weighting Functions for Similarity Measures
CMU Y2 Rosetta GnG Distillation
Jonathan Elsas LTI Student Research Symposium Sept. 14, 2007
Information Organization: Overview
Learning to Rank with Ties
Presentation transcript:

Catching the Drift: Learning Broad Matches from Clickthrough Data Sonal Gupta, Mikhail Bilenko, Matthew Richardson University of Texas at Austin, Microsoft Research

Introduction  Keyword-based online advertising: bidded keywords are extracted from context  Context: query (search ads) or page (content ads)  Broad matching: expanding keywords via keyword-to-keywords mapping  Example: electric cars  tesla, hybrids, toyota prius, golf carts  Broad matching benefits advertisers (increased reach, less campaign tuning), users (more relevant ads), ad platform (higher monetization) Expanded Keywords kw 1 kw 11 kw 12  kw n kw n1 kw n2  Broad Match Expansion Ad Selection and Ranking Ad 1 Ad 2  Ad k Extracted Keywords Keyword Extraction kw 1 kw 2  kw n Query or Web Page Selected Ads

Identifying Broad Matches  Good keyword mappings retrieve relevant ads that users click  How to measure what is relevant and likely to be clicked?  Human judgments: expensive, hard to scale  Past user clicks: provide click data for kw → kw’ when user was shown ad(kw' ) in context of kw  Highly available, less trustworthy  What similarity functions may indicate relevance of kw → kw' ?  Syntactic (edit distance, TF-IDF cosine, string kernels, …)  Co-occurrence (in documents, query sessions, bid campaigns, …)  Expanded representation (search result snippets, category bags, …)

Approach  Task: train a learner to estimate p(click | kw → kw' ) for any kw → kw'  Data  triples from clickthrough logs, where kw → kw' was suggested by previous broad match mappings  Features  Convert each pair to a feature vector capturing similarities etc. (kw → kw') →  For each triple, create an instance: ( ϕ (kw, kw' ), click)  Learner: max-margin averaged perceptron (strong theory, very efficient) ϕ 1 (kw, kw' ) ϕ n (kw, kw' ) … where ϕ i (kw, kw' ) can be any function of kw, kw' or both

5 Example: Creating an Instance  Historical broad match clickthrough data: kw  kw'  ad(kw' )  click event digital slr  canon rebel  Canon Rebel Kit for $499  click seattle baseball  mariners tickets  Mariners season tickets  no click  Feature functions  Instances [ ], 1 [ ], 0 Original kwBroad match kw' ϕ1ϕ1 ϕ2ϕ2 ϕ3ϕ3 digital slrcanon rebel seattle baseballmariners tickets

Experiments  Data  2 months of previous broad match ads from Microsoft Content Ads logs  1 month for training, 1 month for testing  68 features (syntactic, co-occurrence based, etc.); greedy feature selection  Metrics  LogLoss:  LogLoss Lift: difference between obtained LogLoss and an oracle that has access to empirical p(click | kw → kw' ) in test set.  CTR and revenue results in live test with users

Results

Live Test Results  Use CTR prediction to maximize expected revenue  Re-rank mappings to incorporate revenue  +18% revenue, -2% CTR

Online Learning with Amnesia  Advertisers, campaigns, bidded keywords and delivery contexts change very rapidly: high concept drift  Recent data is more informative  Goal: utilize older data while capturing changes in distributions  Averaged Perceptron doesn’t capture drift  Solution: Amnesiac Averaged Perceptron  Exponential weight decay when averaging hypotheses

Results Model-LogLossLogL Lift Prior Feature Selection + Online Learning + Amnesia Online+Feature Selection, No Amnesia Online+Amnesia, No Feature Selection Feature Selection+Amnesia, Weekly Batch

Contributions and Conclusions learning broad matches from implicit feedback  Combining arbitrary similarity measures/features  Using clickthrough logs as implicit feedback  Amnesiac Averaged Perceptron  Exponentially weighted averaging: distant examples “fade out”  Online learning adapts to market dynamics

Thank You!

13 Features and Feature Selection  Co-occurrence feature examples:  User search sessions: keywords searched within 10 mins  Advertiser campaigns: keywords co-bidded by the same advertiser  Past clickthrough rates of original and broad matched keywords  Various syntactic similarities  Various existing broad matching lists  and so on…  Feature Selection:  A total of 68 features  Greedy feature selection

Additional Information  Estimation of expected value of click over all the ads shown for a broad match mapping E(p(click(ad(kw))|q))  Query Expansion vs. Broad Matching  Our broad matching algorithm can be extended for query expansion  But, broad matching is for a fixed set of bidded keywords  Forgetron vs. Amesiac Averaged Perceptron  Forgetron maintains a set of budget support vectors: stores examples explicitly and does not take into account all the data  AAP: weighted average over all the examples, no need to store examples explicitly

Results Model-LogLossLogL Lift Prior Feature Selection + Online Learning + Amnesia Online+Amnesia, No Feature Selection Feature Selection+Amnesia, Weekly Batch Online+Feature Selection, No Amnesia

16 Amnesiac Averaged Perceptron