Locally Constraint Support Vector Clustering

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Introduction to Support Vector Machines (SVM)
CHAPTER 13: Alpaydin: Kernel Machines
Generative Models Thus far we have essentially considered techniques that perform classification indirectly by modeling the training data, optimizing.
Lecture 9 Support Vector Machines
ECG Signal processing (2)
Hierarchical Clustering, DBSCAN The EM Algorithm
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
SVM - Support Vector Machines A new classification method for both linear and nonlinear data It uses a nonlinear mapping to transform the original training.
An Introduction of Support Vector Machine
Classification / Regression Support Vector Machines
Support Vector Machines Instructor Max Welling ICS273A UCIrvine.
An Introduction of Support Vector Machine
Support Vector Machines
Machine learning continued Image source:
Robust Multi-Kernel Classification of Uncertain and Imbalanced Data
ICIP 2000, Vancouver, Canada IVML, ECE, NTUA Face Detection: Is it only for Face Recognition?  A few years earlier  Face Detection Face Recognition 
Pattern Recognition and Machine Learning
Support Vector Machines and Kernel Methods
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
1 Manifold Clustering of Shapes Dragomir Yankov, Eamonn Keogh Dept. of Computer Science & Eng. University of California Riverside.
University of Texas at Austin Machine Learning Group Department of Computer Sciences University of Texas at Austin Support Vector Machines.
Dimensional reduction, PCA
Multiple Human Objects Tracking in Crowded Scenes Yao-Te Tsai, Huang-Chia Shih, and Chung-Lin Huang Dept. of EE, NTHU International Conference on Pattern.
Recovering Articulated Object Models from 3D Range Data Dragomir Anguelov Daphne Koller Hoi-Cheung Pang Praveen Srinivasan Sebastian Thrun Computer Science.
Object Class Recognition Using Discriminative Local Features Gyuri Dorko and Cordelia Schmid.
Support Vector Machines
1 Ensembles of Nearest Neighbor Forecasts Dragomir Yankov, Eamonn Keogh Dept. of Computer Science & Eng. University of California Riverside Dennis DeCoste.
Arizona State University DMML Kernel Methods – Gaussian Processes Presented by Shankar Bhargav.
Jacinto C. Nascimento, Member, IEEE, and Jorge S. Marques
Support Vector Clustering Algorithm presentation by : Jialiang Wu.
Introduction to machine learning
An Introduction to Support Vector Machines Martin Law.
Anomaly detection Problem motivation Machine Learning.
Ch. Eick: Support Vector Machines: The Main Ideas Reading Material Support Vector Machines: 1.Textbook 2. First 3 columns of Smola/Schönkopf article on.
Outline Separating Hyperplanes – Separable Case
Prakash Chockalingam Clemson University Non-Rigid Multi-Modal Object Tracking Using Gaussian Mixture Models Committee Members Dr Stan Birchfield (chair)
CS 8751 ML & KDDSupport Vector Machines1 Support Vector Machines (SVMs) Learning mechanism based on linear programming Chooses a separating plane based.
Non Negative Matrix Factorization
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
Machine Learning Seminar: Support Vector Regression Presented by: Heng Ji 10/08/03.
计算机学院 计算感知 Support Vector Machines. 2 University of Texas at Austin Machine Learning Group 计算感知 计算机学院 Perceptron Revisited: Linear Separators Binary classification.
IEEE TRANSSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE
Classifiers Given a feature representation for images, how do we learn a model for distinguishing features from different classes? Zebra Non-zebra Decision.
An Introduction to Support Vector Machines (M. Law)
Using Support Vector Machines to Enhance the Performance of Bayesian Face Recognition IEEE Transaction on Information Forensics and Security Zhifeng Li,
Presenter : Lin, Shu-Han Authors : Jeen-Shing Wang, Jen-Chieh Chiang
GRASP Learning a Kernel Matrix for Nonlinear Dimensionality Reduction Kilian Q. Weinberger, Fei Sha and Lawrence K. Saul ICML’04 Department of Computer.
Spoken Language Group Chinese Information Processing Lab. Institute of Information Science Academia Sinica, Taipei, Taiwan
Optimal Dimensionality of Metric Space for kNN Classification Wei Zhang, Xiangyang Xue, Zichen Sun Yuefei Guo, and Hong Lu Dept. of Computer Science &
University of Texas at Austin Machine Learning Group Department of Computer Sciences University of Texas at Austin Support Vector Machines.
Radial Basis Function ANN, an alternative to back propagation, uses clustering of examples in the training set.
ACADS-SVMConclusions Introduction CMU-MMAC Unsupervised and weakly-supervised discovery of events in video (and audio) Fernando De la Torre.
Massive Support Vector Regression (via Row and Column Chunking) David R. Musicant and O.L. Mangasarian NIPS 99 Workshop on Learning With Support Vectors.
Present by: Fang-Hui Chu Large Margin Gaussian Mixture Modeling for Phonetic Classification and Recognition Fei Sha*, Lawrence K. Saul University of Pennsylvania.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
1 Kernel Machines A relatively new learning methodology (1992) derived from statistical learning theory. Became famous when it gave accuracy comparable.
A Brief Introduction to Support Vector Machine (SVM) Most slides were from Prof. A. W. Moore, School of Computer Science, Carnegie Mellon University.
SUPPORT VECTOR MACHINES
PREDICT 422: Practical Machine Learning
Support Feature Machine for DNA microarray data
Dept. Computer Science & Engineering, Shanghai Jiao Tong University
Unsupervised Riemannian Clustering of Probability Density Functions
Machine Learning Basics
3.1 Clustering Finding a good clustering of the points is a fundamental issue in computing a representative simplicial complex. Mapper does not place any.
An Introduction to Support Vector Machines
Jianping Fan Dept of CS UNC-Charlotte
SMEM Algorithm for Mixture Models
Presentation transcript:

Locally Constraint Support Vector Clustering Dragomir Yankov, Eamonn Keogh, Kin Fai Kan Computer Science & Eng. Dept. University of California, Riverside

Outline On the need of improving the Support Vector Clustering (SVC) algorithm. Motivation Problem formulation Locally constrained SVC An overview of SVC Applying factor analysis for local outlier detection Regularizing the decision function of SVC Experimental evaluation

Motivation for improving SVC SVC transforms the data in a high dimensional feature space, where a decision function is computed The support-vectors define contours in the original space representing higher density regions The method is theoretically sound and useful for detecting non-convex formations original data detected clusters

Motivation for improving SVC (cont) Parametrizing SVC incorrectly may either disguise some objectively present clusters, or produce multiple unintuitive clusters Correct parametrization is especially hard in the presence of noise (frequently encountered when learning from embedded manifolds) large kernel widths merge the clusters small kernel widths produce multiple unintuitive clusters

Problem formulation How can we make Support Vector Clustering: Less susceptible to noise in the data More resilient to imprecise parametrization

Locally constrained SVC – one class classification Support Vector density estimation Primal formulation Dual formulation

Locally constrained SVC – labeling the closed contours Support Vector Clustering – decision function Labeling the individual classes Build an affinity matrix and find the connected components

Locally constrained SVC – detecting local outliers Factor analysis: Mixture of factor analyzers We can adapt MFA to pinpoint local outliers Points like P1and P2 that deviate a lot from the FA are among the true outliers

Locally constrained SVC – regularizing the decision function To compute the local deviation of each point we use their Mahalanobis distances with respect to the corresponding FA New primal formulation (weighting the slack variables) New dual formulation

Locally constrained SVC – discussion Difference SVC and LSVC Tuning the parameters cannot achieve the same result SVC LSVC SVC tries to accommodate all outliers building complex boundaries SVC SVC Left : SVC tries to accommodate all examples building complex contours and incorrectly bridging the two concentric clusters. Right : LSVC, the proposed here method, detects most outliers. The contours shrink towards the truly dense regions and the two main clusters are separated correctly. Left : SVC for = 8 and = 0.4. Many outliers are now correctly identified, but the rest of the points are split into multiple uninformative clusters. Right : SVC for = 9 and = 0.1. Increasing also cannot achieve the LSVC effect. The contours become very tight and complex and start splitting into multiple clusters Small kernel width detects the outliers but produces multiple unintuitive clusters

Experimental evaluation – synthetic data Gaussian with radial Gaussian distributions LSVC Good parameter values for LSVC are detected automatically. The right clusters are detected SVC SVC is harder to parametrize. The detected clusters are incorrect

Experimental evaluation – synthetic data Swiss roll data with added Gaussian noise LSVC Most of the noise is identified as bounded SVs by LSVC. The correct clusters are detected SVC SVC tends to merge the two large clusters. With supervision the clusters are eventually identified

Experimental evaluation – face images Frey face dataset LSVC LSVC discriminates the two objectively interesting manifolds embedding the data SVC Even with supervision we could not find parameters that separate the two major manifolds with SVC

Experimental evaluation – shape clustering Arrowheads dataset LSVC Some of the classes are similar. There are multiple elements bridging their shape manifolds SVC LSVC achieves 73% accuracy vs 60% for SVC

Conclusion The LSVC method combines both a global and a local view of the data It computes a decision function that defines a global measure of density support MFA complements this with a local view based on the individual analyzers The algorithm improves significantly on the stability of SVC in the presence of noise LSVC allows for easier automatic parameterization of one-class SVMs

All datasets and the code for LSVC can be obtained by writing to the first author: dyankov@cs.ucr.edu THANK YOU!