Mianwei Zhou, Kevin Chen-Chuan Chang University of Illinois at Urbana-Champaign Unifying Learning to Rank and Domain Adaptation -- Enabling Cross-Task.

Slides:



Advertisements
Similar presentations
L3S Research Center University of Hanover Germany
Advertisements

Query Classification Using Asymmetrical Learning Zheng Zhu Birkbeck College, University of London.
Latent Variables Naman Agarwal Michael Nute May 1, 2013.
Mianwei Zhou, Kevin Chen-Chuan Chang University of Illinois at Urbana-Champaign Entity-Centric Document Filtering: Boosting Feature Mapping through Meta-Features.
Diversified Retrieval as Structured Prediction Redundancy, Diversity, and Interdependent Document Relevance (IDR ’09) SIGIR 2009 Workshop Yisong Yue Cornell.
Overview of the TAC2013 Knowledge Base Population Evaluation: Temporal Slot Filling Mihai Surdeanu with a lot help from: Hoa Dang, Joe Ellis, Heng Ji,
Linking Entities in #Microposts ROMIL BANSAL, SANDEEP PANEM, PRIYA RADHAKRISHNAN, MANISH GUPTA, VASUDEVA VARMA INTERNATIONAL INSTITUTE OF INFORMATION TECHNOLOGY,
SIGIR 2013 Recap September 25, 2013.
A Two-Stage Approach to Domain Adaptation for Statistical Classifiers Jing Jiang & ChengXiang Zhai Department of Computer Science University of Illinois.
Query Dependent Pseudo-Relevance Feedback based on Wikipedia SIGIR ‘09 Advisor: Dr. Koh Jia-Ling Speaker: Lin, Yi-Jhen Date: 2010/01/24 1.
Optimizing Estimated Loss Reduction for Active Sampling in Rank Learning Presented by Pinar Donmez joint work with Jaime G. Carbonell Language Technologies.
Learning to Rank: New Techniques and Applications Martin Szummer Microsoft Research Cambridge, UK.
Personalized Search Result Diversification via Structured Learning
 Manmatha MetaSearch R. Manmatha, Center for Intelligent Information Retrieval, Computer Science Department, University of Massachusetts, Amherst.
MANISHA VERMA, VASUDEVA VARMA PATENT SEARCH USING IPC CLASSIFICATION VECTORS.
1 Discovering Unexpected Information from Your Competitor’s Web Sites Bing Liu, Yiming Ma, Philip S. Yu Héctor A. Villa Martínez.
+ Doing More with Less : Student Modeling and Performance Prediction with Reduced Content Models Yun Huang, University of Pittsburgh Yanbo Xu, Carnegie.
Scalable Text Mining with Sparse Generative Models
Finding Advertising Keywords on Web Pages Scott Wen-tau YihJoshua Goodman Microsoft Research Vitor R. Carvalho Carnegie Mellon University.
Adapting Deep RankNet for Personalized Search
Quality-aware Collaborative Question Answering: Methods and Evaluation Maggy Anastasia Suryanto, Ee-Peng Lim Singapore Management University Aixin Sun.
CS344: Introduction to Artificial Intelligence Vishal Vachhani M.Tech, CSE Lecture 34-35: CLIR and Ranking in IR.
Jinhui Tang †, Shuicheng Yan †, Richang Hong †, Guo-Jun Qi ‡, Tat-Seng Chua † † National University of Singapore ‡ University of Illinois at Urbana-Champaign.
1 A Discriminative Approach to Topic- Based Citation Recommendation Jie Tang and Jing Zhang Presented by Pei Li Knowledge Engineering Group, Dept. of Computer.
TREC 2009 Review Lanbo Zhang. 7 tracks Web track Relevance Feedback track (RF) Entity track Blog track Legal track Million Query track (MQ) Chemical IR.
©2008 Srikanth Kallurkar, Quantum Leap Innovations, Inc. All rights reserved. Apollo – Automated Content Management System Srikanth Kallurkar Quantum Leap.
PageRank for Product Image Search Kevin Jing (Googlc IncGVU, College of Computing, Georgia Institute of Technology) Shumeet Baluja (Google Inc.) WWW 2008.
1 Formal Models for Expert Finding on DBLP Bibliography Data Presented by: Hongbo Deng Co-worked with: Irwin King and Michael R. Lyu Department of Computer.
Improving Web Search Ranking by Incorporating User Behavior Information Eugene Agichtein Eric Brill Susan Dumais Microsoft Research.
The Necessity of Combining Adaptation Methods Cognitive Computation Group, University of Illinois Experimental Results Title Ming-Wei Chang, Michael Connor.
 Text Representation & Text Classification for Intelligent Information Retrieval Ning Yu School of Library and Information Science Indiana University.
Xiaoxiao Shi, Qi Liu, Wei Fan, Philip S. Yu, and Ruixin Zhu
Giorgos Giannopoulos (IMIS/”Athena” R.C and NTU Athens, Greece) Theodore Dalamagas (IMIS/”Athena” R.C., Greece) Timos Sellis (IMIS/”Athena” R.C and NTU.
Machine Learning in Ad-hoc IR. Machine Learning for ad hoc IR We’ve looked at methods for ranking documents in IR using factors like –Cosine similarity,
Confidence-Aware Graph Regularization with Heterogeneous Pairwise Features Yuan FangUniversity of Illinois at Urbana-Champaign Bo-June (Paul) HsuMicrosoft.
Contextual Ranking of Keywords Using Click Data Utku Irmak, Vadim von Brzeski, Reiner Kraft Yahoo! Inc ICDE 09’ Datamining session Summarized.
1 A fast algorithm for learning large scale preference relations Vikas C. Raykar and Ramani Duraiswami University of Maryland College Park Balaji Krishnapuram.
Indirect Supervision Protocols for Learning in Natural Language Processing II. Learning by Inventing Binary Labels This work is supported by DARPA funding.
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
Mianwei Zhou, Hongning Wang, Kevin Chen-Chuan Chang University of Illinois Urbana Champaign Learning to Rank from Distant Supervision: Exploiting Noisy.
Biographies, Bollywood, Boom-boxes and Blenders: Domain Adaptation for Sentiment Classification John Blitzer, Mark Dredze and Fernando Pereira University.
Probabilistic Latent Query Analysis for Combining Multiple Retrieval Sources Rong Yan Alexander G. Hauptmann School of Computer Science Carnegie Mellon.
Conceptual structures in modern information retrieval Claudio Carpineto Fondazione Ugo Bordoni
Guest lecture: Feature Selection Alan Qi Dec 2, 2004.
Iterative similarity based adaptation technique for Cross Domain text classification Under: Prof. Amitabha Mukherjee By: Narendra Roy Roll no: Group:
ASSOCIATIVE BROWSING Evaluating 1 Jinyoung Kim / W. Bruce Croft / David Smith for Personal Information.
Learning to Rank: From Pairwise Approach to Listwise Approach Authors: Zhe Cao, Tao Qin, Tie-Yan Liu, Ming-Feng Tsai, and Hang Li Presenter: Davidson Date:
Making Holistic Schema Matching Robust: An Ensemble Approach Bin He Joint work with: Kevin Chen-Chuan Chang Univ. Illinois at Urbana-Champaign.
Toward Entity Retrieval over Structured and Text Data Mayssam Sayyadian, Azadeh Shakery, AnHai Doan, ChengXiang Zhai Department of Computer Science University.
Machine Learning in Practice Lecture 8 Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
UIC at TREC 2006: Blog Track Wei Zhang Clement Yu Department of Computer Science University of Illinois at Chicago.
The Database and Info. Systems Lab. University of Illinois at Urbana-Champaign RankFP : A Framework for Rank Formulation and Processing Hwanjo Yu, Seung-won.
Autumn Web Information retrieval (Web IR) Handout #14: Ranking Based on Click Through data Ali Mohammad Zareh Bidoki ECE Department, Yazd University.
Concept Grounding to Multiple Knowledge Bases via Indirect Supervision
An Empirical Study of Learning to Rank for Entity Search
Restricted Boltzmann Machines for Classification
Deep Web Mining and Learning for Advanced Local Search
Learning to Rank Shubhra kanti karmaker (Santu)
Asymmetric Gradient Boosting with Application to Spam Filtering
Weakly Learning to Match Experts in Online Community
iSRD Spam Review Detection with Imbalanced Data Distributions
MEgo2Vec: Embedding Matched Ego Networks for User Alignment Across Social Networks Jing Zhang+, Bo Chen+, Xianming Wang+, Fengmei Jin+, Hong Chen+, Cuiping.
John Lafferty, Chengxiang Zhai School of Computer Science
Introduction to Machine Reading Comprehension
Measuring Complexity of Web Pages Using Gate
Discriminative Probabilistic Models for Relational Data
Jonathan Elsas LTI Student Research Symposium Sept. 14, 2007
Learning to Rank with Ties
Extracting Why Text Segment from Web Based on Grammar-gram
Presentation transcript:

Mianwei Zhou, Kevin Chen-Chuan Chang University of Illinois at Urbana-Champaign Unifying Learning to Rank and Domain Adaptation -- Enabling Cross-Task Document Scoring

Document Scorer Learning with Queries/Domains 2 Spam DetectionInformation RetrievalSentiment Analysis Handle Various User Needs Queries Tackle Diverse Sources of Data Domains

3 Document Scorer Learning for Different Queries: Learning to Rank Example Application: Entity-Centric Filtering (TREC-KBA 2012) Difficulty: Long Wikipedia pages as queries; Noise keywords in queries. Training Phase Testing Phase Wiki Page RelevantIrrelevant Wiki Page Relevant Irrelevant Wiki Page ? ???

4 Document Scorer Learning for Different Domains: Domain Adaptation I do not like this book, it is very boring... Do not buy this juice extractor, it is leaking... Example Application: Cross-Domain Sentiment Analysis (Blitzer 2006) Difficulty: Different domains use different sentiment keywords. Training Phase Testing Phase It is a very interesting book … Book Reviews This coffee maker has high quality! Kitchen Appliance Reviews

Learning to Rank VS Domain Adaptation 5 Training Phase Testing Phase B Boring Interesting book B Bill Gates Microsoft talk B Chicago Bull Basketball training B Ugly Leaking kitchen Common Challenge Different keyword importance across training and testing phases.

6 Problem: Cross-Task Document Scoring

7 Training Phase Testing Phase B Boring Interesting book B Bill Gates Microsoft talk B Chicago Bull Basketball training B Ugly Leaking kitchen Cross-Task Document Scoring Unify Learning to Rank and Domain Adaptation 1. Task Query (learning to rank) Domain (domain adaptation) 2. Cross-Task: Training and testing phases tackle different tasks.

8 Challenge: Document Scoring across Different Tasks

9 Relevant or Not Document Scoring Principle QueryDocument

10 Document Scoring Principle Q1: Which keywords are important for the query? Q2: How the keywords are contained in the doc? The relevance of a document depends on how it contains keywords that are important for the query. B Bill Gates Microsoft talk Tuesday QueryDocumentKeywords

Requirement of Traditional Learning to Rank: Manually Fulfill the Principle In Feature Design 11 Learning to Rank Models RankSVM RankBoost LambdaMART Abstraction BM25, Language Model, Vector Space, … Output Score Document Relevance

Difficult to Manually Fulfill the Principle for Noisy Query and Complex Document 12 BM25, Language Model,... Q1: Which are important?Q2: How are contained? Ans: High-IDF Ans: TF B Bill Gates Microsoft talk Tuesday Insufficient

Limitation of Traditional Learning to Rank: Leave Heavy Burden to Feature Designers 13 Learning to Rank Models RankSVM RankBoost LambdaMART Leave the Burden to Feature Designers

14 Proposal: Feature Decoupling

Feature Decoupling -- Towards Facilitating the Feature Design 15 Document Scoring Principle How the document contains keywords that are important for the query? Document Scoring Principle How the document contains keywords that are important for the query? Q1: Which keywords are important? Q2: How the keywords are contained?

Feature Decoupling for Entity Centric Document Filtering 16 Intra-Feature B Bill Gates Microsoft talk Tuesday Meta-Feature General: IDF, IsNoun, InEntity,... Structural: PositionInPage, InInfobox, InOpenPara,... Different Position: TFInURL, TFInTitle,... Different Representation: LogTF, NormalizedTF,...

Feature Decoupling for Cross-Domain Sentiment Analysis 17 Intra-FeatureMeta-Feature Different Position: TFInURL, TFInTitle,... Different Representation: LogTF, NormalizedTF,... Good Bad Boring Interesting Tedious High Quality Leaking Broken Pivot Keywords

To Learn Ranker given Decoupled Features, the model should 1. “Recouple” Features; 18 Document RelevanceKeyword Contribution Document Scoring Principle The relevance of a document depends on how it contains keywords that are important for the query? Document Scoring Principle The relevance of a document depends on how it contains keywords that are important for the query?

To Learn Ranker given Decoupled Features, the model should 1. “Recouple” Features; 19 Document Scoring Principle The relevance of a document depends on how it contains keywords that are important for the query? Document Scoring Principle The relevance of a document depends on how it contains keywords that are important for the query? Intra-Feature Meta-Feature Recoupling

20 To Learn Ranker given Decoupled Features, the model should 1. “Recouple” Features; 2. Be Noise Aware. Document Scoring Principle The relevance of a document depends on how it contains keywords that are important for the query? Document Scoring Principle The relevance of a document depends on how it contains keywords that are important for the query? QueryNoisy Keywords listMexico Jeff Noise-Aware

Requirement for Noise-Aware Recoupling: Inferred Sparsity 21 QueryDocumentNoisy Keywords list Mexico Jeff

To Achieve Inferred Sparsity: Two-Layer Scoring Model 22 Keyword Classifier Inferred Sparsity for noisy keywords Contribution for important Keywords

Realize such a Two-Stage Scoring Model is Non-Trivial 23 Keyword Classifier Inferred Sparsity for noisy keywords Contribution for important Keywords

24 Solution: Tree-Structured Restricted Boltzmann Machine

Overview of Tree-Structured Restricted Boltzmann Machine (T-RBM) 25 …

26 … 00

27 … y

28 … 00

Learning Feature Weighting by Likelihood Maximization 29 … Maximize Compute the Gradient by Belief Propagation

30 Experiment

Datasets for Two Different Applications 31 Entity-Centric Document Filtering Dataset: TREC-KBA 29 person entities, 52,238 documents Wikipedia pages as ID pages Cross-Domain Sentiment Analysis Use Dataset released by Blitzer et al. 4 domains, 8000 documents.

T-RBM Outperforms Other Baselines on Both Applications 32 Traditional learning to rank/classify frameworks without Feature Decoupling. Use a simple linear weighting model to combine meta- features and intra-features. Use a boosting framework to combine meta-features and intra-features. Structural Correspondence Learning (SCL), the domain adaptation model proposed by Blitzer et al.

Summary 33 Propose to solve learning to rank and domain adaptation as a unified cross-task document scoring problem. Propose the idea of feature decoupling to facilitate feature design. Propose a noise-aware T-RBM model to “recouple the features”.

34 Thanks! Q&A