LPP-HOG: A New Local Image Descriptor for Fast Human Detection Andy Qing Jun Wang and Ru Bo Zhang IEEE International Symposium.

Slides:



Advertisements
Similar presentations
Introduction to Support Vector Machines (SVM)
Advertisements

Lecture 9 Support Vector Machines
ECG Signal processing (2)
Histograms of Oriented Gradients for Human Detection
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
SVM - Support Vector Machines A new classification method for both linear and nonlinear data It uses a nonlinear mapping to transform the original training.
ONLINE ARABIC HANDWRITING RECOGNITION By George Kour Supervised by Dr. Raid Saabne.
An Introduction of Support Vector Machine
An Introduction of Support Vector Machine
Support Vector Machines
SVM—Support Vector Machines
Support vector machine
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
Machine learning continued Image source:
Computer vision: models, learning and inference Chapter 8 Regression.
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
Groundwater 3D Geological Modeling: Solving as Classification Problem with Support Vector Machine A. Smirnoff, E. Boisvert, S. J.Paradis Earth Sciences.
Computer Vision for Human-Computer InteractionResearch Group, Universität Karlsruhe (TH) cv:hci Dr. Edgar Seemann 1 Computer Vision: Histograms of Oriented.
Histograms of Oriented Gradients for Human Detection Navneet Dalal and Bill Triggs CVPR 2005 Another Descriptor.
Presenter: Hoang, Van Dung
Discriminative and generative methods for bags of features
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
Support Vector Machines (SVMs) Chapter 5 (Duda et al.)
University of Texas at Austin Machine Learning Group Department of Computer Sciences University of Texas at Austin Support Vector Machines.
1 Classification: Definition Given a collection of records (training set ) Each record contains a set of attributes, one of the attributes is the class.
Support Vector Machines Kernel Machines
Sketched Derivation of error bound using VC-dimension (1) Bound our usual PAC expression by the probability that an algorithm has 0 error on the training.
Support Vector Machines
Lecture 10: Support Vector Machines
Greg GrudicIntro AI1 Support Vector Machine (SVM) Classification Greg Grudic.
Classification III Tamara Berg CS Artificial Intelligence Many slides throughout the course adapted from Svetlana Lazebnik, Dan Klein, Stuart Russell,
Support Vector Machine & Image Classification Applications
1 SUPPORT VECTOR MACHINES İsmail GÜNEŞ. 2 What is SVM? A new generation learning system. A new generation learning system. Based on recent advances in.
计算机学院 计算感知 Support Vector Machines. 2 University of Texas at Austin Machine Learning Group 计算感知 计算机学院 Perceptron Revisited: Linear Separators Binary classification.
Kernel Methods A B M Shawkat Ali 1 2 Data Mining ¤ DM or KDD (Knowledge Discovery in Databases) Extracting previously unknown, valid, and actionable.
SVM Support Vector Machines Presented by: Anas Assiri Supervisor Prof. Dr. Mohamed Batouche.
Classifiers Given a feature representation for images, how do we learn a model for distinguishing features from different classes? Zebra Non-zebra Decision.
An Introduction to Support Vector Machines (M. Law)
Recognition II Ali Farhadi. We have talked about Nearest Neighbor Naïve Bayes Logistic Regression Boosting.
Kernels Usman Roshan CS 675 Machine Learning. Feature space representation Consider two classes shown below Data cannot be separated by a hyperplane.
CS 478 – Tools for Machine Learning and Data Mining SVM.
Kernel Methods: Support Vector Machines Maximum Margin Classifiers and Support Vector Machines.
Histograms of Oriented Gradients for Human Detection(HOG)
Support Vector Machines Project מגישים : גיל טל ואורן אגם מנחה : מיקי אלעד נובמבר 1999 הטכניון מכון טכנולוגי לישראל הפקולטה להנדסת חשמל המעבדה לעיבוד וניתוח.
An Introduction to Support Vector Machine (SVM)
CS 1699: Intro to Computer Vision Support Vector Machines Prof. Adriana Kovashka University of Pittsburgh October 29, 2015.
University of Texas at Austin Machine Learning Group Department of Computer Sciences University of Texas at Austin Support Vector Machines.
1  Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Final Exam Review CS479/679 Pattern Recognition Dr. George Bebis 1.
Support Vector Machines Reading: Ben-Hur and Weston, “A User’s Guide to Support Vector Machines” (linked from class web page)
CS 2750: Machine Learning Support Vector Machines Prof. Adriana Kovashka University of Pittsburgh February 17, 2016.
Kernel Methods: Support Vector Machines Maximum Margin Classifiers and Support Vector Machines.
Support Vector Machine: An Introduction. (C) by Yu Hen Hu 2 Linear Hyper-plane Classifier For x in the side of o : w T x + b  0; d = +1; For.
An Introduction of Support Vector Machine In part from of Jinwei Gu.
1 Kernel Machines A relatively new learning methodology (1992) derived from statistical learning theory. Became famous when it gave accuracy comparable.
Roughly overview of Support vector machines Reference: 1.Support vector machines and machine learning on documents. Christopher D. Manning, Prabhakar Raghavan.
A Brief Introduction to Support Vector Machine (SVM) Most slides were from Prof. A. W. Moore, School of Computer Science, Carnegie Mellon University.
Support Vector Machines Reading: Textbook, Chapter 5 Ben-Hur and Weston, A User’s Guide to Support Vector Machines (linked from class web page)
An Introduction of Support Vector Machine Courtesy of Jinwei Gu.
Non-separable SVM's, and non-linear classification using kernels Jakob Verbeek December 16, 2011 Course website:
Support Vector Machines (SVMs) Chapter 5 (Duda et al.) CS479/679 Pattern Recognition Dr. George Bebis.
Support Vector Machine Slides from Andrew Moore and Mingyue Tan.
PREDICT 422: Practical Machine Learning
Support Vector Machine
Cascade for Fast Detection
Geometrical intuition behind the dual problem
Lit part of blue dress and shadowed part of white dress are the same color
Support Vector Machines
Support Vector Machines Introduction to Data Mining, 2nd Edition by
CS 2750: Machine Learning Support Vector Machines
Presentation transcript:

LPP-HOG: A New Local Image Descriptor for Fast Human Detection Andy Qing Jun Wang and Ru Bo Zhang IEEE International Symposium on Knowledge Acquisition and Modeling Workshop, pp Dec. 2008, Wuhan

2 Intelligent Systems Lab. Problem setting Goal: design algorithm for human detection able to perform in real-time Proposed solution: -Use a combination of Histogram of Oriented Gradients (HOG) as a feature vector. - Decrease feature-space dimensionality using Locality Preserving Projection (LPP) - Use Support Vector Machine (SVM) algorithm in reduced feature space to train the classifier

3 Intelligent Systems Lab. HOG general scheme

4 Intelligent Systems Lab. Typical person detection scheme using SVM In practice, effect is very small (about 1%) while some computational time is required* *Navneet Dalal and Bill Triggs. Histograms of Oriented Gradients for Human Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, SanDiego, USA, June Vol. II, pp

5 Intelligent Systems Lab. Computing gradients Mask Type 1D centered 1D uncentered 1D cubic ‑ corrected 2x2 diagonal3x3 Sobel Operator[-1, 0, 1][-1, 1][1, -8, 0, 8, -1] Miss rate at 10 −4 FPPW 11%12.5%12%12.5%14%

6 Intelligent Systems Lab. Accumulate weight votes over spatial cells How many bins should be in histogram? Should we use oriented or non-oriented gradients? How to select weights? Should we use overlapped blocks or not? If yes, then how big should be the overlap? What block size should we use? z

7 Intelligent Systems Lab. Accumulate weight votes over spatial cells How many bins should be in histogram? Should we use oriented or non-oriented gradients? How to select weights? Should we use overlapped blocks or not? If yes, then how big should be the overlap? What block size should we use?

8 Intelligent Systems Lab. Accumulate weight votes over spatial cells How many bins should be in histogram? Should we use oriented or non-oriented gradients? How to select weights? Should we use overlapped blocks or not? If yes, then how big should be the overlap? What block size should we use?

9 Intelligent Systems Lab. Contrast normalization - L2-norm followed by clipping (limiting the maximum values of v to 0.2) and renormalising

10 Intelligent Systems Lab. Making feature vector Variants of HOG descriptors. (a) A rectangular HOG (R-HOG) descriptor with 3 × 3 blocks of cells. (b) Circular HOG (C-HOG) descriptor with the central cell divided into angular sectors as in shape contexts. (c) A C-HOG descriptor with a single central cell.

11 Intelligent Systems Lab. HOG feature vector for one block AngleMagnitude Binary voting Magnitude voting Feature vector extends while window moves

12 Intelligent Systems Lab. HOG example In each triplet: (1) the input image, (2) the corresponding R-HOG feature vector (only the dominant orientation of each cell is shown), (3) the dominant orientations selected by the SVM (obtained by multiplying the feature vector by the corresponding weights from the linear SVM).

13 Intelligent Systems Lab. Support Vector Machine (SVM)

14 Intelligent Systems Lab. Problem setting for SVM x1x1 x2x2 w T x + b = 0 w T x + b < 0 w T x + b > 0 A hyper-plane in the feature space (Unit-length) normal vector of the hyper-plane: n denotes +1 denotes -1

15 Intelligent Systems Lab. x1x1 x2x2 How would you classify these points using a linear discriminant function in order to minimize the error rate? denotes +1 denotes -1 Infinite number of answers! Which one is the best? Problem setting for SVM

16 Intelligent Systems Lab. Large Margin Linear Classifier We know that The margin width is: x1x1 x2x2 denotes +1 denotes -1 Margin w T x + b = 0 w T x + b = -1 w T x + b = 1 x+x+ x+x+ x-x- n Support Vectors

17 Intelligent Systems Lab. Large Margin Linear Classifier Formulation: x1x1 x2x2 denotes +1 denotes -1 Margin w T x + b = 0 w T x + b = -1 w T x + b = 1 x+x+ x+x+ x-x- n such that

18 Intelligent Systems Lab. Large Margin Linear Classifier Formulation: x1x1 x2x2 denotes +1 denotes -1 Margin w T x + b = 0 w T x + b = -1 w T x + b = 1 x+x+ x+x+ x-x- n such that

19 Intelligent Systems Lab. Large Margin Linear Classifier Formulation: x1x1 x2x2 denotes +1 denotes -1 Margin w T x + b = 0 w T x + b = -1 w T x + b = 1 x+x+ x+x+ x-x- n such that

20 Intelligent Systems Lab. Solving the Optimization Problem s.t. Quadratic programming with linear constraints s.t. Lagrangian Function

21 Intelligent Systems Lab. Solving the Optimization Problem s.t.

22 Intelligent Systems Lab. Solving the Optimization Problem s.t., and Lagrangian Dual Problem

23 Intelligent Systems Lab. Solving the Optimization Problem The solution has the form: From KKT condition, we know: Thus, only support vectors have x1x1 x2x2 w T x + b = 0 w T x + b = -1 w T x + b = 1 x+x+ x+x+ x-x- Support Vectors

24 Intelligent Systems Lab. Solving the Optimization Problem The linear discriminant function is: Notice it relies on a dot product between the test point x and the support v ectors x i Also keep in mind that solving the optimization problem involved computing the dot products x i T x j between all pairs of training points

25 Intelligent Systems Lab. Large Margin Linear Classifier What if data is not linear separable? (noisy data, outliers, etc.) Slack variables ξ i can be added to allow miss-classification of dif ficult or noisy data points x1x1 x2x2 denotes +1 denotes -1 w T x + b = 0 w T x + b = -1 w T x + b = 1

26 Intelligent Systems Lab. Large Margin Linear Classifier Formulation: such that Parameter C can be viewed as a way to control over-fitting.

27 Intelligent Systems Lab. Large Margin Linear Classifier Formulation: (Lagrangian Dual Problem) such that

28 Intelligent Systems Lab. Datasets that are linearly separable with noise work out great: 0 x 0 x x2x2 0 x But what are we going to do if the dataset is just too hard? How about … mapping data to a higher-dimensional space: Non-linear SVMs

29 Intelligent Systems Lab. General idea: the original input space can be mapped to some higher-dime nsional feature space where the training set is separable: Φ: x → φ(x) Non-linear SVMs: Feature Space

30 Intelligent Systems Lab. With this mapping, our discriminant function is now: No need to know this mapping explicitly, because we only use the dot product o f feature vectors in both the training and test. A kernel function is defined as a function that corresponds to a dot product of two feature vectors in some expanded feature space: Non-linear SVMs: The Kernel Trick

31 Intelligent Systems Lab. Linear kernel: Examples of commonly-used kernel functions: Polynomial kernel: Gaussian (Radial-Basis Function (RBF) ) kernel: Non-linear SVMs: The Kernel Trick

32 Intelligent Systems Lab. Nonlinear SVM: Optimization Formulation: (Lagrangian Dual Problem) such that The solution of the discriminant function is The optimization technique is the same.

33 Intelligent Systems Lab. Support Vector Machine: Algorithm 1. Choose a kernel function 2. Choose a value for C 3. Solve the quadratic programming problem (many algorithms and software packages available) 4. Construct the discriminant function from the support vectors

34 Intelligent Systems Lab. Summary: Support Vector Machine 1. Large Margin Classifier Better generalization ability & less over-fitting 2. The Kernel Trick Map data points to higher dimensional space in order to make them linearly separable. Since only dot product is used, we do not need to represent the mapping explicitly.

35 Intelligent Systems Lab. Back to the proposed paper

36 Intelligent Systems Lab. Proposed algorithm parameters - Bins in histogram: 8 - Cell size: 4x4 pixels - Block size: 2x2 cells (8x8 pixels) - Image size: 64x128 pixels (8x16 blocks) - Feature vector size: 2x2x8x8x16=4096

37 Intelligent Systems Lab. LPP Algorithm Main idea: find matrix which will project original data into a space with lower dimensionality while preserving similarity between data (data which are close to each other in original space should be close after projection)

38 Intelligent Systems Lab. LPP Algorithm Is it correct? Add constraints Can be represented as a generalized eigenvalue problem Is it correct? By selecting d smallest eigenvalues and corresponding eigenvectors dimensionality reduction is achieved

39 Intelligent Systems Lab. Solving different scale problem

40 Intelligent Systems Lab. Some results Dimension d Detection rate PCA-HOG features (labeled’ *’) vs LPP-HOG features (labeled ˅ ’) Detection example

41 Intelligent Systems Lab. Conclusions - Fast human detection algorithm based on HOG features is presented - no information about computational speed is given - Proposed method is similar to PCA-HOG - feature space dimensionality decreased using LPP - why do we need to make LPP instead of finding eigenvectors from original feature space? - some equations seems to be wrong - Reference papers are very few Navneet Dalal “Finding People in Images and Videos” PhD Thesis. Institut National Polytechnique de Grenoble / INRIA Grenoble, Grenoble, July Navneet Dalal and Bill Triggs, “Histograms of Oriented Gradients for Human Detection”. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, SanDiego, USA, June Vol. II, pp Paisitkriangkrai, S., Shen, C. and Zhang, J. “Performance evaluation of local features in human classification and detection”, IET Computer Vision, vol.2, issue 4, pp ,December 2008