Final Exam Review CS479/679 Pattern Recognition Dr. George Bebis 1.

Slides:



Advertisements
Similar presentations
Introduction to Support Vector Machines (SVM)
Advertisements

Generative Models Thus far we have essentially considered techniques that perform classification indirectly by modeling the training data, optimizing.
ECG Signal processing (2)
Support Vector Machine & Its Applications Mingyue Tan The University of British Columbia Nov 26, 2004 A portion (1/3) of the slides are taken from Prof.
SVM - Support Vector Machines A new classification method for both linear and nonlinear data It uses a nonlinear mapping to transform the original training.
Classification / Regression Support Vector Machines
Pattern Recognition and Machine Learning
An Introduction of Support Vector Machine
Support Vector Machines and Kernels Adapted from slides by Tim Oates Cognition, Robotics, and Learning (CORAL) Lab University of Maryland Baltimore County.
Support Vector Machines
SVM—Support Vector Machines
Support vector machine
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
Lecture 14 – Neural Networks
Support Vector Machines (and Kernel Methods in general)
Support Vector Machines and Kernel Methods
Fuzzy Support Vector Machines (FSVMs) Weijia Wang, Huanren Zhang, Vijendra Purohit, Aditi Gupta.
Support Vector Machines (SVMs) Chapter 5 (Duda et al.)
Prénom Nom Document Analysis: Linear Discrimination Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Announcements See Chapter 5 of Duda, Hart, and Stork. Tutorial by Burge linked to on web page. “Learning quickly when irrelevant attributes abound,” by.
Announcements  Project proposal is due on 03/11  Three seminars this Friday (EB 3105) Dealing with Indefinite Representations in Pattern Recognition.
Support Vector Machines Based on Burges (1998), Scholkopf (1998), Cristianini and Shawe-Taylor (2000), and Hastie et al. (2001) David Madigan.
Support Vector Machine (SVM) Classification
Discriminant Functions Alexandros Potamianos Dept of ECE, Tech. Univ. of Crete Fall
Lecture outline Support vector machines. Support Vector Machines Find a linear hyperplane (decision boundary) that will separate the data.
Lecture 10: Support Vector Machines
SVM (Support Vector Machines) Base on statistical learning theory choose the kernel before the learning process.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Expectation-Maximization (EM) Chapter 3 (Duda et al.) – Section 3.9
Linear Discriminant Functions Chapter 5 (Duda et al.)
Greg GrudicIntro AI1 Support Vector Machine (SVM) Classification Greg Grudic.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Classification III Tamara Berg CS Artificial Intelligence Many slides throughout the course adapted from Svetlana Lazebnik, Dan Klein, Stuart Russell,
Ch. Eick: Support Vector Machines: The Main Ideas Reading Material Support Vector Machines: 1.Textbook 2. First 3 columns of Smola/Schönkopf article on.
This week: overview on pattern recognition (related to machine learning)
Based on: The Nature of Statistical Learning Theory by V. Vapnick 2009 Presentation by John DiMona and some slides based on lectures given by Professor.
Support Vector Machine & Image Classification Applications
CS 8751 ML & KDDSupport Vector Machines1 Support Vector Machines (SVMs) Learning mechanism based on linear programming Chooses a separating plane based.
Kernel Methods A B M Shawkat Ali 1 2 Data Mining ¤ DM or KDD (Knowledge Discovery in Databases) Extracting previously unknown, valid, and actionable.
SVM Support Vector Machines Presented by: Anas Assiri Supervisor Prof. Dr. Mohamed Batouche.
CS 478 – Tools for Machine Learning and Data Mining SVM.
Kernel Methods: Support Vector Machines Maximum Margin Classifiers and Support Vector Machines.
Ohad Hageby IDC Support Vector Machines & Kernel Machines IP Seminar 2008 IDC Herzliya.
Linear hyperplanes as classifiers Usman Roshan. Hyperplane separators.
Linear Models for Classification
University of Texas at Austin Machine Learning Group Department of Computer Sciences University of Texas at Austin Support Vector Machines.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
CSE4334/5334 DATA MINING CSE4334/5334 Data Mining, Fall 2014 Department of Computer Science and Engineering, University of Texas at Arlington Chengkai.
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Part 9: Review.
Linear hyperplanes as classifiers Usman Roshan. Hyperplane separators.
Greg GrudicIntro AI1 Support Vector Machine (SVM) Classification Greg Grudic.
Kernel Methods: Support Vector Machines Maximum Margin Classifiers and Support Vector Machines.
SVMs in a Nutshell.
1 Kernel Machines A relatively new learning methodology (1992) derived from statistical learning theory. Became famous when it gave accuracy comparable.
A Brief Introduction to Support Vector Machine (SVM) Most slides were from Prof. A. W. Moore, School of Computer Science, Carnegie Mellon University.
Linear Discriminant Functions Chapter 5 (Duda et al.) CS479/679 Pattern Recognition Dr. George Bebis.
Linear machines márc Decison surfaces We focus now on the decision surfaces Linear machines = linear decision surface Non-optimal solution but.
Support Vector Machines (SVMs) Chapter 5 (Duda et al.) CS479/679 Pattern Recognition Dr. George Bebis.
Neural networks and support vector machines
CS 9633 Machine Learning Support Vector Machines
PREDICT 422: Practical Machine Learning
LECTURE 16: SUPPORT VECTOR MACHINES
LINEAR AND NON-LINEAR CLASSIFICATION USING SVM and KERNELS
Support Vector Machines Introduction to Data Mining, 2nd Edition by
Pattern Recognition CS479/679 Pattern Recognition Dr. George Bebis
Linear machines 28/02/2017.
LECTURE 17: SUPPORT VECTOR MACHINES
Support Vector Machines and Kernels
Presentation transcript:

Final Exam Review CS479/679 Pattern Recognition Dr. George Bebis 1

Final Exam Material Midterm Exam Material Linear Discriminant Functions Support Vector Machines Expectation-Maximization Algorithm 2 Case studies are also included in the final exam

Linear Discriminant Functions General form of linear discriminant: What is the form of the decision boundary? What is the meaning of w and w 0 ? – The decision boundary is a hyperplane ; its orientation is determined by w and its location by w 0. 3

Linear Discriminant Functions What does g(x) measure? – Distance of x from the decision boundary (hyperplane) 4

Linear Discriminant Functions How do we find w and w 0 ? – Apply learning using a set of labeled training examples What is the effect of each training example? – Places a constraint on the solution 5 feature space (y 1, y 2 ) a1a1a1a1 a2a2a2a2 solution space ( ɑ 1, ɑ 2 )

Linear Discriminant Functions Iterative optimization – what is the main idea? – Minimize some error function J(α) iteratively 6 (k) α(k) (k+1) α(k+1) search direction learning rate

Linear Discriminant Functions Gradient descent method Newton method Perceptron rule 7

Support Vector Machines What is the capacity of a classifier? What is the VC dimension of a classifier? What is structural risk minimization? – Find solutions that (1) minimize the empirical risk and (2) have low VC dimension. – It can be shown that: 8 with probability (1-δ)

Support Vector Machines What is the margin of separation? How is it defined? What is the relationship between VC dimension and margin of separation? – VC dimension is minimized by maximizing the margin of separation. 9 support vectors

Support Vector Machines What is the criterion being optimized by SVMs? 10 maximize margin:

Support Vector Machines SVM solution depends only on the support vectors: Soft margin classifier – tolerate “outliers” 11

Support Vector Machines Non-linear SVM – what is the main idea? – Map data to a high dimensional space h 12

Support Vector Machines What is the kernel trick? – Compute dot products using a kernel function 13 K(x,y)=(x. y) d e.g., polynomial kernel:

Support Vector Machines Important comments about SVMs – SVM is based on exact optimization (no local optima). – Its complexity depends on the number of support vectors, not on the dimensionality of the transformed space. – Performance depends on the choice of the kernel and its parameters. 14

Expectation-Maximization (EM) What is the EM algorithm? – An iterative method to perform ML estimation i.e., max p(D/ θ) When is EM useful? – Most useful for problems where the data is incomplete or can be thought as being incomplete. 15

Expectation-Maximization (EM) What are the steps of the EM algorithm? – Initialization: θ 0 – Expectation Step: – Maximization Step: – Test for convergence: Convergence properties of EM ? – Solution depends on the initial estimate θ 0 – No guarantee to find global maximum but stable 16

Expectation-Maximization (EM) What is a mixture of Gaussians? How are the parameters of MoGs estimated? – Using the EM algorithm What is the main idea behind using EM for estimating the MoGs parameters? – Introduce “hidden variables: 17

Expectation-Maximization (EM) Explain the EM steps for MoGs 18

Expectation-Maximization (EM) 19 Explain the EM steps for MoGs