November 10, 2004Dmitriy Fradkin, CIKM'041 A Design Space Approach to Analysis of Information Retrieval Adaptive Filtering Systems Dmitriy Fradkin, Paul.

Slides:



Advertisements
Similar presentations
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Advertisements

Improvements and extras Paul Thomas CSIRO. Overview of the lectures 1.Introduction to information retrieval (IR) 2.Ranked retrieval 3.Probabilistic retrieval.
1 Evaluation Rong Jin. 2 Evaluation  Evaluation is key to building effective and efficient search engines usually carried out in controlled experiments.
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
Explorations in Tag Suggestion and Query Expansion Jian Wang and Brian D. Davison Lehigh University, USA SSM 2008 (Workshop on Search in Social Media)
Evaluating Search Engine
K nearest neighbor and Rocchio algorithm
Morris LeBlanc.  Why Image Retrieval is Hard?  Problems with Image Retrieval  Support Vector Machines  Active Learning  Image Processing ◦ Texture.
Carnegie Mellon 1 Maximum Likelihood Estimation for Information Thresholding Yi Zhang & Jamie Callan Carnegie Mellon University
1/ 30. Problems for classical IR models Introduction & Background(LSI,SVD,..etc) Example Standard query method Analysis standard query method Seeking.
1 LM Approaches to Filtering Richard Schwartz, BBN LM/IR ARDA 2002 September 11-12, 2002 UMASS.
Exploration & Exploitation in Adaptive Filtering Based on Bayesian Active Learning Yi Zhang, Jamie Callan Carnegie Mellon Univ. Wei Xu NEC Lab America.
Online Learning for Web Query Generation: Finding Documents Matching a Minority Concept on the Web Rayid Ghani Accenture Technology Labs, USA Rosie Jones.
Affinity Rank Yi Liu, Benyu Zhang, Zheng Chen MSRA.
WXGB6106 INFORMATION RETRIEVAL Week 3 RETRIEVAL EVALUATION.
The Relevance Model  A distribution over terms, given information need I, (Lavrenko and Croft 2001). For term r, P(I) can be dropped w/o affecting the.
Search and Retrieval: Relevance and Evaluation Prof. Marti Hearst SIMS 202, Lecture 20.
1 Probabilistic Language-Model Based Document Retrieval.
Masquerade Detection Mark Stamp 1Masquerade Detection.
APPLICATIONS OF DATA MINING IN INFORMATION RETRIEVAL.
Processing of large document collections Part 2 (Text categorization) Helena Ahonen-Myka Spring 2006.
IR Evaluation Evaluate what? –user satisfaction on specific task –speed –presentation (interface) issue –etc. My focus today: –comparative performance.
©2008 Srikanth Kallurkar, Quantum Leap Innovations, Inc. All rights reserved. Apollo – Automated Content Management System Srikanth Kallurkar Quantum Leap.
Learning to Classify Short and Sparse Text & Web with Hidden Topics from Large- scale Data Collections Xuan-Hieu PhanLe-Minh NguyenSusumu Horiguchi GSIS,
1 Formal Models for Expert Finding on DBLP Bibliography Data Presented by: Hongbo Deng Co-worked with: Irwin King and Michael R. Lyu Department of Computer.
A Simple Unsupervised Query Categorizer for Web Search Engines Prashant Ullegaddi and Vasudeva Varma Search and Information Extraction Lab Language Technologies.
Mining the Web to Create Minority Language Corpora Rayid Ghani Accenture Technology Labs - Research Rosie Jones Carnegie Mellon University Dunja Mladenic.
Latent Semantic Analysis Hongning Wang Recap: vector space model Represent both doc and query by concept vectors – Each concept defines one dimension.
Relevance Feedback Hongning Wang What we have learned so far Information Retrieval User results Query Rep Doc Rep (Index) Ranker.
Binxing Jiao et. al (SIGIR ’10) Presenter : Lin, Yi-Jhen Advisor: Dr. Koh. Jia-ling Date: 2011/4/25 VISUAL SUMMARIZATION OF WEB PAGES.
Greedy is not Enough: An Efficient Batch Mode Active Learning Algorithm Chen, Yi-wen( 陳憶文 ) Graduate Institute of Computer Science & Information Engineering.
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
A Model for Learning the Semantics of Pictures V. Lavrenko, R. Manmatha, J. Jeon Center for Intelligent Information Retrieval Computer Science Department,
1 Using The Past To Score The Present: Extending Term Weighting Models with Revision History Analysis CIKM’10 Advisor : Jia Ling, Koh Speaker : SHENG HONG,
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
Powerpoint Templates Page 1 Powerpoint Templates Scalable Text Classification with Sparse Generative Modeling Antti PuurulaWaikato University.
Learning from Positive and Unlabeled Examples Investigator: Bing Liu, Computer Science Prime Grant Support: National Science Foundation Problem Statement.
Project by: Cirill Aizenberg, Dima Altshuler Supervisor: Erez Berkovich.
A Repetition Based Measure for Verification of Text Collections and for Text Categorization Dmitry V.Khmelev Department of Mathematics, University of Toronto.
Techniques for Collaboration in Text Filtering 1 Ian Soboroff Department of Computer Science and Electrical Engineering University of Maryland, Baltimore.
Modern information retreival Chapter. 02: Modeling (Latent Semantic Indexing)
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Supervised Learning Resources: AG: Conditional Maximum Likelihood DP:
Mining Dependency Relations for Query Expansion in Passage Retrieval Renxu Sun, Chai-Huat Ong, Tat-Seng Chua National University of Singapore SIGIR2006.
KNN & Naïve Bayes Hongning Wang Today’s lecture Instance-based classifiers – k nearest neighbors – Non-parametric learning algorithm Model-based.
Iterative similarity based adaptation technique for Cross Domain text classification Under: Prof. Amitabha Mukherjee By: Narendra Roy Roll no: Group:
Chapter 15 The Chi-Square Statistic: Tests for Goodness of Fit and Independence PowerPoint Lecture Slides Essentials of Statistics for the Behavioral.
Active Feedback in Ad Hoc IR Xuehua Shen, ChengXiang Zhai Department of Computer Science University of Illinois, Urbana-Champaign.
Flexible Automatic Motion Blending with Registration Curves
1 Adaptive Subjective Triggers for Opinionated Document Retrieval (WSDM 09’) Kazuhiro Seki, Kuniaki Uehara Date: 11/02/09 Speaker: Hsu, Yu-Wen Advisor:
Date: 2012/5/28 Source: Alexander Kotov. al(CIKM’11) Advisor: Jia-ling, Koh Speaker: Jiun Jia, Chiou Interactive Sense Feedback for Difficult Queries.
Carnegie Mellon School of Computer Science Language Technologies Institute CMU Team-1 in TDT 2004 Workshop 1 CMU TEAM-A in TDT 2004 Topic Tracking Yiming.
Relevance Feedback Hongning Wang
The Effect of Database Size Distribution on Resource Selection Algorithms Luo Si and Jamie Callan School of Computer Science Carnegie Mellon University.
Usefulness of Quality Click- through Data for Training Craig Macdonald, ladh Ounis Department of Computing Science University of Glasgow, Scotland, UK.
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008 Annotations by Michael L. Nelson.
More on HMMs and Multiple Sequence Alignment BMI/CS 776 Mark Craven March 2002.
Relevant Document Distribution Estimation Method for Resource Selection Luo Si and Jamie Callan School of Computer Science Carnegie Mellon University
Introduction to Information Retrieval Introduction to Information Retrieval Lecture 15: Text Classification & Naive Bayes 1.
KNN & Naïve Bayes Hongning Wang
Queensland University of Technology
Evaluation of IR Systems
MMS Software Deliverables: Year 1
Object Localization Goal: detect the location of an object within an image Fully supervised: Training data labeled with object category and ground truth.
John Lafferty, Chengxiang Zhai School of Computer Science
A Potpurri of topics Paul Kantor
Authors: Wai Lam and Kon Fan Low Announcer: Kyu-Baek Hwang
Relevance and Reinforcement in Interactive Browsing
Retrieval Performance Evaluation - Measures
INFORMATION RETRIEVAL TECHNIQUES BY DR. ADNAN ABID
Presentation transcript:

November 10, 2004Dmitriy Fradkin, CIKM'041 A Design Space Approach to Analysis of Information Retrieval Adaptive Filtering Systems Dmitriy Fradkin, Paul Kantor DIMACS, Rutgers University

November 10, 2004Dmitriy Fradkin, CIKM'042 What Is This Work About? Small-scale view: We analyze differences between two implementations of Rocchio method and discuss choices of parameters. Large-scale view: The problem of constructing an IR/AF system can be seen as an optimization problem in a large design space. (Well-known methods are simply points in this space.)

November 10, 2004Dmitriy Fradkin, CIKM'043 Large-Scale View Use optimization methods to find optimal choices of parameters. These optimal choices do not have to correspond to well-known methods or standard practices. Design space optimization methods have been suggested for designing VLSI chips [Bahuman et. al. 2002], airplanes [Schwabacher and Gelsey, 1996; Zha et. al ] and HVAC systems [Szykman 1997].

November 10, 2004Dmitriy Fradkin, CIKM'044 What’s in a name? We find that even a single “name” involves an enormous number of design choices. TREC2002 Adaptive Filtering –DIMACS: Rocchio method –Chinese Academy of Sciences: Rocchio Method One method performs almost twice as well as the other.

November 10, 2004Dmitriy Fradkin, CIKM'045 For any system: Choose Data Representation Construct Initial Classifier Training Phase: Incorporate labeled examples Supplement with “pseudo positives” and “pseudo negatives” Set the threshold Filtering Phase: as new documents arrive Evaluate performance Update the classifier model Update threshold

November 10, 2004Dmitriy Fradkin, CIKM'046 All of these are usually: Characterized informally, as a choice, and the exclusion of alternatives. Seen as points on a map – but to understand the significance of these choices we need to explore the real territory. So: we must interpolate between the choices made in one method and those made in another.

November 10, 2004Dmitriy Fradkin, CIKM'047 Interpolation Identify the corresponding design decisions Develop a “path” between them –sometimes called a “homotopy” from the topological concept of smoothly distorting one shape (say a coffee cup) into another (say, a doughnut). Study the effectiveness along various paths among design options.

November 10, 2004Dmitriy Fradkin, CIKM'048 Interpolation Aspects for IR/AF Term Representation Term Weighting Computing Scores Setting Classifier Threshold Document Set Representation Pseudolabeled Documents in Training

November 10, 2004Dmitriy Fradkin, CIKM'049 Interpolation Aspects (cont.) Query Initialization Unjudged document in test Query Update Quitting Strategy

November 10, 2004Dmitriy Fradkin, CIKM'0410 Example: Term Representation Where f’(t,d) is number of times a term occurs in a document

November 10, 2004Dmitriy Fradkin, CIKM'0411 Example: Term Weighting DIMACS: CAS: Homotopy: i’(t) is the number of documents, in training set T, containing term t.

November 10, 2004Dmitriy Fradkin, CIKM'0412 Example: Score Computation DIMACS: CAS: Homotopy: i’(t) is the number of documents, in training set T, containing term t. W is a diagonal matrix of weights

November 10, 2004Dmitriy Fradkin, CIKM'0413 Example: Score Interpolation Same mapping for scores and for thresholds from CAS scale to DIMACS scale: Homotopy:

November 10, 2004Dmitriy Fradkin, CIKM'0414 Example: Setting Thresholds DIMACS: CAS: Homotopy: is chosen to optimize utility Threshold for query q after seeing document i:

November 10, 2004Dmitriy Fradkin, CIKM'0415 Example: Set Representation DIMACS CAS Homotopy

November 10, 2004Dmitriy Fradkin, CIKM'0416 Example: Pseudo-labeled Documents CAS method does not make use of pseudo-labeled documents in training stage DIMACS method: Given “density” parameters (d+ and d-) and “proportion” (p+ and p-), score unlabeled training documents and choose top and bottom sets according to “proportion”. Then pick documents out of these sets according to corresponding “density”. Interpolate between density and proportion parameters (DIMACS) and 0 (CAS).

November 10, 2004Dmitriy Fradkin, CIKM'0417 Example: Query Initialization General Formula: DIMACS: CAS: Homotopy:

November 10, 2004Dmitriy Fradkin, CIKM'0418 Example: Unjudged Documents A submitted document for which there is no label is “unjudged”. DIMACS ignores such documents. CAS considers such documents pseudo-negative if its score is less than 0.6. Can view this as a threshold:

November 10, 2004Dmitriy Fradkin, CIKM'0419 Example: Query Update General Formula: DIMACS: CAS: Homotopy:

November 10, 2004Dmitriy Fradkin, CIKM'0420 Example: Quitting Strategy DIMACS: if after 50 submissions the utility is negative, stop submitting for this topic CAS: no quitting strategy Alternatively:

November 10, 2004Dmitriy Fradkin, CIKM'0421 Experimental Evaluation TREC11 Data - Reuters Corpus v1 23,000 training; 800,000 test 100 topics (50 assessor, 50 intersection) 3 positive and 0 negative examples per topic T+ - all positive documents; D+ - submitted positive; D- - submitted negative; Du – submitted unlabelled

November 10, 2004Dmitriy Fradkin, CIKM'0422 Diagonal Interpolation

November 10, 2004Dmitriy Fradkin, CIKM'0423 Documents Retrieved

November 10, 2004Dmitriy Fradkin, CIKM'0424 Parameter Analysis It is possible to analyze effect of individual parameters at each point in space by taking “small steps” along the parameter axis. Requires a lot of computational effort Results may not be easy to interpret

November 10, 2004Dmitriy Fradkin, CIKM'0425 Example of Parameter Analysis Effect of individual parameters on number of relevant and nonrelevant documents retrieved around 0.8 point

November 10, 2004Dmitriy Fradkin, CIKM'0426 Results based on topic type Comparison of CAS results and 0.8 diagonal homotopy point

November 10, 2004Dmitriy Fradkin, CIKM'0427 Additional Experiments Reordered TREC documents Experimented with 77 topics on OHSUMED dataset ( as training data, as test) The results are similar to those on the original TREC task.

November 10, 2004Dmitriy Fradkin, CIKM'0428 Result of Experiments with Reordering Lambda Average T11SU Standard Deviation Average Results on 5 re-orderings of TREC test set:

November 10, 2004Dmitriy Fradkin, CIKM'0429 OHSUMED Results

November 10, 2004Dmitriy Fradkin, CIKM'0430 Documents Retrieved: OHSUMED

November 10, 2004Dmitriy Fradkin, CIKM'0431 Discussion We demonstrate the design complexity hidden under “Rocchio method” We provide specific models for interpolating between design choices These interpolation options can work for methods that are significantly more different (for example Rocchio and SVM).

November 10, 2004Dmitriy Fradkin, CIKM'0432 Discussion (cont.) These models should help researchers explore their systems, and regions “between systems” Suggests a new approach to designing IR systems: finding a set of (interpolation) parameters optimizing performance This can be done with existing optimization methods.

November 10, 2004Dmitriy Fradkin, CIKM'0433 A Note on Interpolation Limits The need for two endpoint systems is not very restrictive: Some interpolation parameters can be moved beyond [0,1] interval. The endpoints themselves can be moved.

November 10, 2004Dmitriy Fradkin, CIKM'0434 Abstract Interpolation More abstractly: do not interpolate every single parameter –work at higher abstraction levels Ex: representation block, scoring block, thresholding block, etc. Can use this with several systems This is at a lower level than ensembles of classifiers.

November 10, 2004Dmitriy Fradkin, CIKM'0435 Caveat In moving to large design space we still face two major problems: The range of parameters cannot be explored exhaustively, and non-smooth optimization is needed Requires a lot of labeled data that is usually produced manually and is in short supply.

November 10, 2004Dmitriy Fradkin, CIKM'0436 Acknowledgments KD-D group via NSF grant EIA Andrei Anghelescu, Vladimir Menkov Jamie Callan Members of DIMACS MMS project CAS researchers Ian Soboroff Anonymous reviewers