Support vector machines

Slides:



Advertisements
Similar presentations
3.6 Support Vector Machines
Advertisements

Introduction to Support Vector Machines (SVM)
Support Vector Machine
Regularized risk minimization
Lecture 9 Support Vector Machines
ECG Signal processing (2)
An Introduction of Support Vector Machine
Classification / Regression Support Vector Machines
Support Vector Machines Instructor Max Welling ICS273A UCIrvine.
1 Lecture 5 Support Vector Machines Large-margin linear classifier Non-separable case The Kernel trick.
Support vector machine
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
Separating Hyperplanes
Support Vector Machines
Support Vector Machines (and Kernel Methods in general)
Classification and risk prediction Usman Roshan. Disease risk prediction What is the best method to predict disease risk? –We looked at the maximum likelihood.
Support Vector Classification (Linearly Separable Case, Primal) The hyperplanethat solves the minimization problem: realizes the maximal margin hyperplane.
Classification Problem 2-Category Linearly Separable Case A- A+ Malignant Benign.
Support Vector Machine (SVM) Classification
Support Vector Machines and Kernel Methods
1 Computational Learning Theory and Kernel Methods Tianyi Jiang March 8, 2004.
SVM Support Vectors Machines
Lecture 10: Support Vector Machines
Greg GrudicIntro AI1 Support Vector Machine (SVM) Classification Greg Grudic.
Optimization Theory Primal Optimization Problem subject to: Primal Optimal Value:
Support Vector Machines
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
Linear hyperplanes as classifiers Usman Roshan. Hyperplane separators.
CS Statistical Machine learning Lecture 18 Yuan (Alan) Qi Purdue CS Oct
Computational Intelligence: Methods and Applications Lecture 23 Logistic discrimination and support vectors Włodzisław Duch Dept. of Informatics, UMK Google:
Kernels Usman Roshan CS 675 Machine Learning. Feature space representation Consider two classes shown below Data cannot be separated by a hyperplane.
CS 478 – Tools for Machine Learning and Data Mining SVM.
Kernel Methods: Support Vector Machines Maximum Margin Classifiers and Support Vector Machines.
Linear hyperplanes as classifiers Usman Roshan. Hyperplane separators.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Support Vector Machines
1  Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Support Vector Machines Tao Department of computer science University of Illinois.
Text Classification using Support Vector Machine Debapriyo Majumdar Information Retrieval – Spring 2015 Indian Statistical Institute Kolkata.
Linear hyperplanes as classifiers Usman Roshan. Hyperplane separators.
Greg GrudicIntro AI1 Support Vector Machine (SVM) Classification Greg Grudic.
Computational Intelligence: Methods and Applications Lecture 24 SVM in the non-linear case Włodzisław Duch Dept. of Informatics, UMK Google: W Duch.
Kernel Methods: Support Vector Machines Maximum Margin Classifiers and Support Vector Machines.
Support Vector Machine: An Introduction. (C) by Yu Hen Hu 2 Linear Hyper-plane Classifier For x in the side of o : w T x + b  0; d = +1; For.
Generalization Error of pac Model  Let be a set of training examples chosen i.i.d. according to  Treat the generalization error as a r.v. depending on.
An Introduction of Support Vector Machine Courtesy of Jinwei Gu.
1 Support Vector Machines: Maximum Margin Classifiers Machine Learning and Pattern Recognition: September 23, 2010 Piotr Mirowski Based on slides by Sumit.
Non-separable SVM's, and non-linear classification using kernels Jakob Verbeek December 16, 2011 Course website:
Support vector machines
CS 9633 Machine Learning Support Vector Machines
PREDICT 422: Practical Machine Learning
Support Vector Machine
LINEAR CLASSIFIERS The Problem: Consider a two class task with ω1, ω2.
Large Margin classifiers
Computational Intelligence: Methods and Applications
Geometrical intuition behind the dual problem
Support Vector Machines
Regularized risk minimization
Jan Rupnik Jozef Stefan Institute
An Introduction to Support Vector Machines
Kernels Usman Roshan.
Statistical Learning Dong Liu Dept. EEIS, USTC.
CSSE463: Image Recognition Day 14
Support vector machines
Support Vector Machines
Usman Roshan CS 675 Machine Learning
Support vector machines
Support vector machines
SVMs for Document Ranking
Support Vector Machines 2
Presentation transcript:

Support vector machines Usman Roshan

Separating hyperplanes For two sets of points there are many hyperplane separators Which one should we choose for classification? In other words which one is most likely to produce least error?

Margin of a plane We define the margin as the minimum distance to training points (distance to closest point) y x

Margin error bound Margin error theorem (7.3 from Learning with kernels)

Maximum margin hyperplane Previous theorem dictates that the best hyperplane is the one that maximizes the minimum distance of all training points to the plane (Learning with kernels, Scholkopf and Smola, 2002) also called the margin. Its expected error is at most the fraction of misclassified points plus a complexity term (Learning with kernels, Scholkopf and Smola, 2002)

Optimally separating hyperplane How do we find the optimally separating (maximum margin) hyperplane? Recall distance of a point to the plane defined earlier

Hyperplane separators   r xp x w

Distance of a point to the separating plane And so the distance to the plane r is given by or where y is -1 if the point is on the left side of the plane and +1 otherwise.

Support vector machine: optimally separating hyperplane Distance of point x (with label y) to the hyperplane is given by We want this to be at least some value By scaling w we can obtain infinite solutions. Therefore we require that So we minimize ||w|| to maximize the distance which gives us the SVM optimization problem.

Support vector machine: optimally separating hyperplane SVM optimization criterion (primal form) We can solve this with Lagrange multipliers to reveal the dual optimization problem.

SVM dual problem Let L be the Lagrangian: Solving dL/dw=0 and dL/dw0=0 gives The xi for which αi is non-zero are called support vectors. Substituting w into the Lagrangian gives us the dual optimization problem

Support vector machine: optimally separating hyperplane

Another look at the SVM objective Consider the objective: This is non-convex and harder to solve than the convex SVM objective Roughly it measures the total sum of distances of points to the plane. Misclassified points are given a negative distance whereas correct ones have a 0 distance.

SVM objective The SVM objective is Compare this with The SVM objective can be viewed as a convex approximation of this by separating the numerator and denominator The SVM objective is equivalent to minimizing the regularized hinge risk:

Hinge loss empirical risk Hinge risk: But the max function is non-differentiable. Therefore we use the sub-gradient:

Hinge loss regularized risk How does the hinge loss gradient below change for the regularized risk?

Least squares vs hinge risk

Inseparable case and 0/1 loss What is there is no separating hyperplane? For example XOR function. One solution: consider all hyperplanes and select the one with the minimal number of misclassified points This is known as the 0/1 loss optimization problem Turns out to be NP-complete and even NP-complete to polynomially approximate (Learning with kernels, Scholkopf and Smola, and references on website) Thus hard to solve in practice but is less sensitive to outliers than least squares and hinge

Inseparable case If we measure error as roughly the sum of the distance of misclassified points to the plane then we can still solve for a support vector machine in polynomial time Note that total distance error can be considerably larger than number of misclassified points.

Optimally separating hyperplane with errors       w x

SVM software Plenty of SVM software out there. Two popular packages: SVM-light LIBSVM

Kernels What if no separating hyperplane exists (non-linear data)? Consider the XOR function. In a higher dimensional space we can find a separating hyperplane

Kernels Recall that the solution to the SVM is obtained by applying KKT rules (a generalization of Lagrange multipliers). The problem to solve becomes

Kernels The dot product can be replaced by a matrix K(i,j)=xiTxj or a positive definite matrix K.

Kernels With the kernel approach we can avoid explicit calculation of features in high dimensions How do we find the best kernel? Multiple Kernel Learning (MKL) solves it for K as a linear combination of base kernels.