Presentation is loading. Please wait.

Presentation is loading. Please wait.

Support Vector Machine (SVM) Classification

Similar presentations


Presentation on theme: "Support Vector Machine (SVM) Classification"— Presentation transcript:

1 Support Vector Machine (SVM) Classification
Greg Grudic Greg Grudic Intro AI

2 Last Class Linear separating hyperplanes for binary classification
Rosenblatt’s Perceptron Algorithm Based on Gradient Descent Convergence theoretically guaranteed if data is linearly separable Infinite number of solutions For nonlinear data: Mapping data into a nonlinear space where it is linearly separable (or almost) However, convergence still not guaranteed… Greg Grudic Intro AI

3 Questions? Greg Grudic Intro AI

4 Why Classification? Uncertainty world Signals Sensing Actions
Computation Not typically addressed in CS State Symbols (The Grounding Problem) Decisions/Planning Agent Greg Grudic Introduction to AI

5 The Problem Domain for Project Test 1: Identifying (and Navigating) Paths
Non-path Data Data Construct a Classifier Classifier Data Path labeled Image 4/17/2017 Intro AI

6 Today’s Lecture Goals Support Vector Machine (SVM) Classification
Another algorithm for linear separating hyperplanes A Good text on SVMs: Bernhard Schölkopf and Alex Smola. Learning with Kernels. MIT Press, Cambridge, MA, 2002 Greg Grudic Intro AI

7 Support Vector Machine (SVM) Classification
Classification as a problem of finding optimal (canonical) linear hyperplanes. Optimal Linear Separating Hyperplanes: In Input Space In Kernel Space Can be non-linear Greg Grudic Intro AI

8 Linear Separating Hyper-Planes
How many lines can separate these points? Which line should we use? NO! Greg Grudic Intro AI

9 Initial Assumption: Linearly Separable Data
Greg Grudic Intro AI

10 Linear Separating Hyper-Planes
Greg Grudic Intro AI

11 Linear Separating Hyper-Planes
Given data: Finding a separating hyperplane can be posed as a constraint satisfaction problem (CSP): Or, equivalently: If data is linearly separable, there are an infinite number of hyperplanes that satisfy this CSP Greg Grudic Intro AI

12 The Margin of a Classifier
Take any hyper-plane (P0) that separates the data Put a parallel hyper-plane (P1) on a point in class 1 closest to P0 Put a second parallel hyper-plane (P2) on a point in class -1 closest to P0 The margin (M) is the perpendicular distance between P1 and P2 Greg Grudic Intro AI

13 Calculating the Margin of a Classifier
P2 P0: Any separating hyperplane P1: Parallel to P0, passing through closest point in one class P2: Parallel to P0, passing through point closest to the opposite class P0 P1 Margin (M): distance measured along a line perpendicular to P1 and P2 Greg Grudic Intro AI

14 SVM Constraints on the Model Parameters
Model parameters must be chosen such that, for on P1 and for on P2: For any P0, these constraints are always attainable. Given the above, then the linear separating boundary lies half way between P1 and P2 and is given by: Resulting Classifier: Greg Grudic Intro AI

15 Remember: signed distance from a point to a hyperplane:
Hyperplane define by: Greg Grudic Intro AI

16 Calculating the Margin (1)
Greg Grudic Intro AI

17 Calculating the Margin (2)
Signed Distance Take absolute value to get the unsigned margin: Greg Grudic Intro AI

18 Different P0’s have Different Margins
P0: Any separating hyperplane P1: Parallel to P0, passing through closest point in one class P2: Parallel to P0, passing through point closest to the opposite class P0 P1 Margin (M): distance measured along a line perpendicular to P1 and P2 Greg Grudic Intro AI

19 Different P0’s have Different Margins
P0: Any separating hyperplane P1: Parallel to P0, passing through closest point in one class P2: Parallel to P0, passing through point closest to the opposite class P0 P1 Margin (M): distance measured along a line perpendicular to P1 and P2 Greg Grudic Intro AI

20 Different P0’s have Different Margins
P0: Any separating hyperplane P1: Parallel to P0, passing through closest point in one class P2: Parallel to P0, passing through point closest to the opposite class P2 P0 P1 Margin (M): distance measured along a line perpendicular to P1 and P2 Greg Grudic Intro AI

21 How Do SVMs Choose the Optimal Separating Hyperplane (boundary)?
Find the that maximizes the margin! P1 Margin (M): distance measured along a line perpendicular to P1 and P2 Greg Grudic Intro AI

22 SVM: Constraint Optimization Problem
Given data: Minimize subject to: The Lagrange Function Formulation is used to solve this Minimization Problem Greg Grudic Intro AI

23 The Lagrange Function Formulation
For every constraint we introduce a Lagrange Multiplier: The Lagrangian is then defined by: Where - the primal variables are - the dual variables are Goal: Minimize Lagrangian w.r.t. primal variables, and Maximize w.r.t. dual variables Intro AI Greg Grudic

24 Derivation of the Dual Problem
At the saddle point (extremum w.r.t. primal) This give the conditions Substitute into to get the dual problem Greg Grudic Intro AI

25 Using the Lagrange Function Formulation, we get the Dual Problem
Maximize Subject to Greg Grudic Intro AI

26 Properties of the Dual Problem
Solving the Dual gives a solution to the original constraint optimization problem For SVMs, the Dual problem is a Quadratic Optimization Problem which has a globally optimal solution Gives insights into the NON-Linear formulation for SVMs Greg Grudic Intro AI

27 Support Vector Expansion (1)
is also computed from the optimal dual variables Greg Grudic Intro AI

28 Support Vector Expansion (2)
Substitute OR Greg Grudic Intro AI

29 What are the Support Vectors?
Maximized Margin Greg Grudic Intro AI

30 Why do we want a model with only a few SVs?
Leaving out an example that does not become an SV gives the same solution! Theorem (Vapnik and Chervonenkis, 1974): Let be the number of SVs obtained by training on N examples randomly drawn from P(X,Y), and E be an expectation. Then Greg Grudic Intro AI

31 Next Class Nonlinear data Now a quiz! Greg Grudic Intro AI


Download ppt "Support Vector Machine (SVM) Classification"

Similar presentations


Ads by Google