Dept. of Computer Science & Engineering, CUHK Pseudo Relevance Feedback with Biased Support Vector Machine in Multimedia Retrieval Steven C.H. Hoi 14-Oct,

Slides:



Advertisements
Similar presentations
Query Classification Using Asymmetrical Learning Zheng Zhu Birkbeck College, University of London.
Advertisements

Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Introduction to Support Vector Machines (SVM)
Support Vector Machine
SVM - Support Vector Machines A new classification method for both linear and nonlinear data It uses a nonlinear mapping to transform the original training.
Classification / Regression Support Vector Machines
ICONIP 2005 Improve Naïve Bayesian Classifier by Discriminative Training Kaizhu Huang, Zhangbing Zhou, Irwin King, Michael R. Lyu Oct
Pattern Recognition and Machine Learning
Support Vector Machines
SVM—Support Vector Machines
Support vector machine
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
Optimal Design Laboratory | University of Michigan, Ann Arbor 2011 Design Preference Elicitation Using Efficient Global Optimization Yi Ren Panos Y. Papalambros.
Software Quality Ranking: Bringing Order to Software Modules in Testing Fei Xing Michael R. Lyu Ping Guo.
Fei Xing1, Ping Guo1,2 and Michael R. Lyu2
Relevance Feedback Content-Based Image Retrieval Using Query Distribution Estimation Based on Maximum Entropy Principle Irwin King and Zhong Jin Nov
Content Based Image Clustering and Image Retrieval Using Multiple Instance Learning Using Multiple Instance Learning Xin Chen Advisor: Chengcui Zhang Department.
Support Vector Machines (and Kernel Methods in general)
Fuzzy Support Vector Machines (FSVMs) Weijia Wang, Huanren Zhang, Vijendra Purohit, Aditi Gupta.
1 Learning on Relevance Feedback in Content-based Image Retrieval Hoi, Chu-Hong ( Steven ) Supervisor: Prof. Michael R. Lyu Venue:RM1027 Date: 11:00a.m.
University of Texas at Austin Machine Learning Group Department of Computer Sciences University of Texas at Austin Support Vector Machines.
Prénom Nom Document Analysis: Linear Discrimination Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
ACM Multimedia th Annual Conference, October , 2004
An Introduction to Kernel-Based Learning Algorithms K.-R. Muller, S. Mika, G. Ratsch, K. Tsuda and B. Scholkopf Presented by: Joanna Giforos CS8980: Topics.
Relevance Feedback based on Parameter Estimation of Target Distribution K. C. Sia and Irwin King Department of Computer Science & Engineering The Chinese.
SVM Active Learning with Application to Image Retrieval
1 Integrating User Feedback Log into Relevance Feedback by Coupled SVM for Content-Based Image Retrieval 9-April, 2005 Steven C. H. Hoi *, Michael R. Lyu.
Supervised Distance Metric Learning Presented at CMU’s Computer Vision Misc-Read Reading Group May 9, 2007 by Tomasz Malisiewicz.
Support Vector Machines and Kernel Methods
Support Vector Machines
Presentation in IJCNN 2004 Biased Support Vector Machine for Relevance Feedback in Image Retrieval Hoi, Chu-Hong Steven Department of Computer Science.
Semi-Supervised Clustering Jieping Ye Department of Computer Science and Engineering Arizona State University
Presented by Zeehasham Rasheed
Bing LiuCS Department, UIC1 Learning from Positive and Unlabeled Examples Bing Liu Department of Computer Science University of Illinois at Chicago Joint.
Relevance Feedback Content-Based Image Retrieval Using Query Distribution Estimation Based on Maximum Entropy Principle Irwin King and Zhong Jin The Chinese.
Text Classification With Labeled and Unlabeled Data Presenter: Aleksandar Milisic Supervisor: Dr. David Albrecht.
Optimizing Learning with SVM Constraint for Content-based Image Retrieval* Steven C.H. Hoi 1th March, 2004 *Note: The copyright of the presentation material.
Active Learning for Class Imbalance Problem
A Simple Unsupervised Query Categorizer for Web Search Engines Prashant Ullegaddi and Vasudeva Varma Search and Information Extraction Lab Language Technologies.
Glasgow 02/02/04 NN k networks for content-based image retrieval Daniel Heesch.
计算机学院 计算感知 Support Vector Machines. 2 University of Texas at Austin Machine Learning Group 计算感知 计算机学院 Perceptron Revisited: Linear Separators Binary classification.
10/18/ Support Vector MachinesM.W. Mak Support Vector Machines 1. Introduction to SVMs 2. Linear SVMs 3. Non-linear SVMs References: 1. S.Y. Kung,
Partially Supervised Classification of Text Documents by Bing Liu, Philip Yu, and Xiaoli Li Presented by: Rick Knowles 7 April 2005.
ICML2004, Banff, Alberta, Canada Learning Larger Margin Machine Locally and Globally Kaizhu Huang Haiqin Yang, Irwin King, Michael.
Using Support Vector Machines to Enhance the Performance of Bayesian Face Recognition IEEE Transaction on Information Forensics and Security Zhifeng Li,
Greedy is not Enough: An Efficient Batch Mode Active Learning Algorithm Chen, Yi-wen( 陳憶文 ) Graduate Institute of Computer Science & Information Engineering.
PSEUDO-RELEVANCE FEEDBACK FOR MULTIMEDIA RETRIEVAL Seo Seok Jun.
Competence Centre on Information Extraction and Image Understanding for Earth Observation 29th March 2007 Category - based Semantic Search Engine 1 Mihai.
Gang WangDerek HoiemDavid Forsyth. INTRODUCTION APROACH (implement detail) EXPERIMENTS CONCLUSION.
SemiBoost : Boosting for Semi-supervised Learning Pavan Kumar Mallapragada, Student Member, IEEE, Rong Jin, Member, IEEE, Anil K. Jain, Fellow, IEEE, and.
Bing LiuCS Department, UIC1 Chapter 8: Semi-supervised learning.
Design of PCA and SVM based face recognition system for intelligent robots Department of Electrical Engineering, Southern Taiwan University, Tainan County,
Semi-Automatic Image Annotation Liu Wenyin, Susan Dumais, Yanfeng Sun, HongJiang Zhang, Mary Czerwinski and Brent Field Microsoft Research.
Data Mining, ICDM '08. Eighth IEEE International Conference on Duy-Dinh Le National Institute of Informatics Hitotsubashi, Chiyoda-ku Tokyo,
Support Vector Machines
Jen-Tzung Chien, Meng-Sung Wu Minimum Rank Error Language Modeling.
26/01/20161Gianluca Demartini Ranking Categories for Faceted Search Gianluca Demartini L3S Research Seminars Hannover, 09 June 2006.
11 A Classification-based Approach to Question Routing in Community Question Answering Tom Chao Zhou 1, Michael R. Lyu 1, Irwin King 1,2 1 The Chinese.
Virtual Examples for Text Classification with Support Vector Machines Manabu Sassano Proceedings of the 2003 Conference on Emprical Methods in Natural.
Feature Selction for SVMs J. Weston et al., NIPS 2000 오장민 (2000/01/04) Second reference : Mark A. Holl, Correlation-based Feature Selection for Machine.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Support Vector Machines Reading: Ben-Hur and Weston, “A User’s Guide to Support Vector Machines” (linked from class web page)
Greg GrudicIntro AI1 Support Vector Machine (SVM) Classification Greg Grudic.
University Of Seoul Ubiquitous Sensor Network Lab Query Dependent Pseudo-Relevance Feedback based on Wikipedia 전자전기컴퓨터공학 부 USN 연구실 G
High resolution product by SVM. L’Aquila experience and prospects for the validation site R. Anniballe DIET- Sapienza University of Rome.
Semi-Supervised Learning Using Label Mean
Semi-Supervised Clustering
LINEAR AND NON-LINEAR CLASSIFICATION USING SVM and KERNELS
Concave Minimization for Support Vector Machine Classifiers
SVMs for Document Ranking
Presentation transcript:

Dept. of Computer Science & Engineering, CUHK Pseudo Relevance Feedback with Biased Support Vector Machine in Multimedia Retrieval Steven C.H. Hoi 14-Oct, 2003

Outline Introduction Related Work and Motivation SVM and one-class SVM Biased Support Vector Machine (BSVM) Pseudo Relevance Feedback (PRF) with BSVM Experimental results Future Work and Conclusions

Introduction Relevance Feedback (RF) –Is an interactive technique for query reformulation. –Has been shown successfully used in CBIR. Drawbacks of RF –Need user ’ s intervention to provide relevance information. This impose a burden to users. –Most of RF algorithms focus only on positive examples. Pseudo RF (PRF) alleviate the user ’ s burden –Successfully used in text information retrieval. –Use potential positive examples and negative examples to refine the retrieval quality.

Related Work Classical RF Algorithms –From heuristic weighting adjustment to optimal Learning, Transductive Learning (EM, D-EM), DA, SVM, 1-SVM Motivation of BSVM –Classical RF algorithms have limited help in the context of PRF. –Biased Feature in PRF: potential positive examples are limited and clustered in some way, while much negative information is available and scattered in different ways. –In Ref. [1], Yan ec al. propose use SVM-ensembles to shape PRF. –SVMs have shown good generalization performance for classification tasks. –However, classical SVMs treat positive and negative examples symmetrically. –We proposed the Biased SVM to attack the biased feature in PRF.

SVM & 1-SVM SVM –implements the principle of structural risk minimization and can provide good generalization performance. –In the linearly separable case, the goal of SVM is to look for a pair of hyperplanes which give the maximum margin.

1-SVM –1-SVM only focuses on estimating the density of positive examples. –Two kinds of dierent formulation of 1-SVM in the literature. –The sphere-based approach

BSVM –The strategy of our technique is that we take two sphere- hyperplanes to describe the data. –We hope the inner sphere- hyperplane can be as small as possible while it includes most of the positive data. –On the other hand, we want the outer sphere-hyperplane to be as large as possible while it contains least of the negative data.

BSVM Let us consider the following training data: where n is the number of observations, m is the dimension of the input space, and Y is the class label of the training data. The objective function for finding the optimal sphere-hyperplane can be formulated below:

The optimization task can be solved by Lagrange multipliers:

Let us take the partial derivative of L with respect to the parameters respectively. Then setting their partial derivatives to 0, we obtain the following equations,

From the equations, we can we obtain the dual form of the primal optimization problem: The decision function can be constructed as follows,

Differences between BSVM and SVM –Geometric interpretation Compared with the linear decision hyperplane in SVM, BSVM takes a sphere to cluster the class of positive examples, while another class for negative examples is pushed away from the sphere. –Mathematic Perspective Constraint in optimization of SVM: While it becomes another form in BSVM: This means the positive class C 1 will receive more weight than the negative class C 2. It will make the final decision boundary biased toward to the positive class.

PRF with BSVM Compared with SVM –SVM will classify many unseen samples as the positive class while some of them actually do not belong to the class. –BSVM will be more reasonable.

Compared with 1-SVM –Without incorporating the negative information, one-class SVM will classify many negative samples to the positive class. –BSVM can model the data well with the negative constraints.

PRF with BSVM Implementation –Figuring out the dual optimization equation (Equation (*)) belongs to a well-know problem called quadratic programming (QP) problem.Equation (*) –After solving the QP problem, we can obtain a set of optimal parameters ‘ s. Then, we can compute the evaluation function for each image as follows,

Experiment Results Application to CBIR –Using the Corel Image Feature Dataset –Around 68,000 features in the dataset –Selected Features: Color Histogram (32) and Co-occurrence texture (16). –Experiment Settings For an initial query, Top N ranking images are selected as positive. Last M ranking image are selected as negative. N << M, their values are determined by a training set. –Three PRF algorithms are compared: SVM, 1-SVM and BSVM based PRF

Experiments Evaluation Metrics –AP (Average Precision) –AR (Average Recall) Experimental Results –AP Curves

AR Curves

Discussions BSVM based PRF algorithm is effective in improving the retrieval performance of CBIR system and achieves better performance than other two classical approaches. Note that our improvement do not need any intervention from users. However, we may need to look for more sophisticated algorithms for selecting the suitable values of N and M in our PRF scheme.

Future Work With application to Video Retrieval (VR) –Our proposed scheme can be extended to used in VR. –More Challenging for RF in VR –Multimodal PRF algorithm will be investigated: Text + Visual + Audio. Learning Algorithms –Partial Labels and Semi-Supervised Learning Labeled training set is usually a small subset of in multimedia database. Unlabeled working sets are important clues for improve the retrieval performance in Multimedia retrieval. In literature, Bayesian SVC, Semi-SVM and constrained EM are hot topics in Machine Learning. Can we combine the supervised and unsupervised learning for attacking the multimedia retrieval problem? For example, combining the constrained EM algorithm and discriminative classifiers, such as SVM, MPM, etc.?

Conclusions We investigate the pseudo-relevance feedback techniques in multimedia retrieval. We propose and formulate a novel PRF technique with biased support vector machine in image retrieval. Experimental results demonstrate that the BSVM-based PRF algorithm is effective and promising for improving the retrieval performance in CBIR and Multimedia retrieval.

FAQ Thanks!

References [1] R. Yan, et al., “ Multimedia search with Pseudo Relevance Feedback ”, CIVR2003, LNCS2728, Springer, [2] T.S. Huang and X.S. Zhou. Image retrieval by relevance feedback: from heuristic weight adjustment to optimal learning methods. In Proc. IEEE International Conference on Image Processing (ICIP'01), Thessaloniki, Greece, October [3] C.H. Hoi, K. Huang, M.R. Lyu, I. King, “ Pseudo relevance feedback with Biased Support Vector Machine in Image Retrieval ”, submitted to ECCV2004.