Project Overview CSE 6367 – Computer Vision Vassilis Athitsos University of Texas at Arlington.

Slides:



Advertisements
Similar presentations
Statistical Learning of Multi-View Face Detection
Advertisements

Rapid Object Detection using a Boosted Cascade of Simple Features Paul Viola, Michael Jones Conference on Computer Vision and Pattern Recognition 2001.
Rapid Object Detection using a Boosted Cascade of Simple Features Paul Viola, Michael Jones Conference on Computer Vision and Pattern Recognition 2001.
Face detection Behold a state-of-the-art face detector! (Courtesy Boris Babenko)Boris Babenko.
AdaBoost & Its Applications
Face detection Many slides adapted from P. Viola.
Cos 429: Face Detection (Part 2) Viola-Jones and AdaBoost Guest Instructor: Andras Ferencz (Your Regular Instructor: Fei-Fei Li) Thanks to Fei-Fei Li,
EE462 MLCV Lecture 5-6 Object Detection – Boosting Tae-Kyun Kim.
The Viola/Jones Face Detector Prepared with figures taken from “Robust real-time object detection” CRL 2001/01, February 2001.
The Viola/Jones Face Detector (2001)
HCI Final Project Robust Real Time Face Detection Paul Viola, Michael Jones, Robust Real-Time Face Detetion, International Journal of Computer Vision,
Rapid Object Detection using a Boosted Cascade of Simple Features
Robust Real-time Object Detection by Paul Viola and Michael Jones ICCV 2001 Workshop on Statistical and Computation Theories of Vision Presentation by.
Ensemble Tracking Shai Avidan IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE February 2007.
Face detection and recognition Many slides adapted from K. Grauman and D. Lowe.
Learning and Vision: Discriminative Models
Adaboost and its application
Robust Real-Time Object Detection Paul Viola & Michael Jones.
Viola and Jones Object Detector Ruxandra Paun EE/CS/CNS Presentation
Fast Human Detection Using a Novel Boosted Cascading Structure With Meta Stages Yu-Ting Chen and Chu-Song Chen, Member, IEEE.
Computer Vision CSPP Artificial Intelligence March 3, 2004.
Foundations of Computer Vision Rapid object / face detection using a Boosted Cascade of Simple features Presented by Christos Stoilas Rapid object / face.
F ACE D ETECTION FOR A CCESS C ONTROL By Dmitri De Klerk Supervisor: James Connan.
Face Detection CSE 576. Face detection State-of-the-art face detection demo (Courtesy Boris Babenko)Boris Babenko.
FACE DETECTION AND RECOGNITION By: Paranjith Singh Lohiya Ravi Babu Lavu.
Kullback-Leibler Boosting Ce Liu, Hueng-Yeung Shum Microsoft Research Asia CVPR 2003 Presented by Derek Hoiem.
Face Detection using the Viola-Jones Method
Using Statistic-based Boosting Cascade Weilong Yang, Wei Song, Zhigang Qiao, Michael Fang 1.
Introduction to Computer Vision Olac Fuentes Computer Science Department University of Texas at El Paso El Paso, TX, U.S.A.
Detecting Pedestrians Using Patterns of Motion and Appearance Paul Viola Microsoft Research Irfan Ullah Dept. of Info. and Comm. Engr. Myongji University.
Rectangle Filters and AdaBoost CSE 6367 – Computer Vision Vassilis Athitsos University of Texas at Arlington.
Learning a Fast Emulator of a Binary Decision Process Center for Machine Perception Czech Technical University, Prague ACCV 2007, Tokyo, Japan Jan Šochman.
Window-based models for generic object detection Mei-Chen Yeh 04/24/2012.
Sign Classification Boosted Cascade of Classifiers using University of Southern California Thang Dinh Eunyoung Kim
Lecture 29: Face Detection Revisited CS4670 / 5670: Computer Vision Noah Snavely.
Face detection Slides adapted Grauman & Liebe’s tutorial
Visual Object Recognition
DIEGO AGUIRRE COMPUTER VISION INTRODUCTION 1. QUESTION What is Computer Vision? 2.
Terrorists Team members: Ágnes Bartha György Kovács Imre Hajagos Wojciech Zyla.
Robust Real-time Face Detection by Paul Viola and Michael Jones, 2002 Presentation by Kostantina Palla & Alfredo Kalaitzis School of Informatics University.
ECE738 Advanced Image Processing Face Detection IEEE Trans. PAMI, July 1997.
Tracking CSE 6367 – Computer Vision Vassilis Athitsos University of Texas at Arlington.
Face Detection Ying Wu Electrical and Computer Engineering Northwestern University, Evanston, IL
Fast Similarity Search in Image Databases CSE 6367 – Computer Vision Vassilis Athitsos University of Texas at Arlington.
Robust Real Time Face Detection
Adaboost and Object Detection Xu and Arun. Principle of Adaboost Three cobblers with their wits combined equal Zhuge Liang the master mind. Failure is.
Lecture 09 03/01/2012 Shai Avidan הבהרה: החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
The Viola/Jones Face Detector A “paradigmatic” method for real-time object detection Training is slow, but detection is very fast Key ideas Integral images.
Bibek Jang Karki. Outline Integral Image Representation of image in summation format AdaBoost Ranking of features Combining best features to form strong.
Learning to Detect Faces A Large-Scale Application of Machine Learning (This material is not in the text: for further information see the paper by P.
Hand Gesture Recognition Using Haar-Like Features and a Stochastic Context-Free Grammar IEEE 高裕凱 陳思安.
FACE DETECTION : AMIT BHAMARE. WHAT IS FACE DETECTION ? Face detection is computer based technology which detect the face in digital image. Trivial task.
Face detection Behold a state-of-the-art face detector! (Courtesy Boris Babenko)Boris Babenko slides adapted from Svetlana Lazebnik.
CS-498 Computer Vision Week 9, Class 2 and Week 10, Class 1
A Brief Introduction on Face Detection Mei-Chen Yeh 04/06/2010 P. Viola and M. J. Jones, Robust Real-Time Face Detection, IJCV 2004.
Face detection Many slides adapted from P. Viola.
Hand Detection with a Cascade of Boosted Classifiers Using Haar-like Features Qing Chen Discover Lab, SITE, University of Ottawa May 2, 2006.
AdaBoost Algorithm and its Application on Object Detection Fayin Li.
1 Munther Abualkibash University of Bridgeport, CT.
Reading: R. Schapire, A brief introduction to boosting
2. Skin - color filtering.
Cascade for Fast Detection
Session 7: Face Detection (cont.)
Error Correcting Code.
Presented by Minh Hoai Nguyen Date: 28 March 2007
Lit part of blue dress and shadowed part of white dress are the same color
Cos 429: Face Detection (Part 2) Viola-Jones and AdaBoost Guest Instructor: Andras Ferencz (Your Regular Instructor: Fei-Fei Li) Thanks to Fei-Fei.
ADABOOST(Adaptative Boosting)
Adaboost for faces. Material
Lecture 29: Face Detection Revisited
Presentation transcript:

Project Overview CSE 6367 – Computer Vision Vassilis Athitsos University of Texas at Arlington

Project Task: implementing a face detector. Components: –AdaBoost. –Skin detection. –Bootstrapping. –Cascades.

Training Data Positive examples: –Images in training_faces. Negative examples: –Images in training_nonfaces. What windows should be used? –The whole image, or subwindows? –You will have to choose. How much training data should be used? –Too few lead to less accuracy. –Too many may not fit in memory, may lead to slow training.

Test Data test_cropped_faces: cropped images of faces. –You know there is a face, and you know where it is. test_face_photos: –You know there is a face, you don’t know where it is (but you can annotate that). test_nonfaces: –You know there is no face.

Reporting Performance

False positives: –How many nonface windows were classified as faces? False negatives: –How many faces were classified as nonfaces? Efficiency: –How much time does it take to process all test images (in the three test directories)?

Using Skin Color In color images, faces have skin color. Skin detection is much faster than face detection. You should use skin detection to make the detector faster AND more accurate. –You should compare results obtained using skin detection and without using skin detection. –Choices must be justified based on training data. Avoid setting arbitrary thresholds. If setting thresholds, consider how they work on training data to determine threshold values.

Choosing Negative Examples How many training examples does a photograph generate? –Assume that the photograph does not contain faces.

Choosing Negative Examples How many training examples does a photograph generate? –Assume that the photograph does not contain faces. Every position and scale defines a training example. –Too many examples to use for AdaBoost training. Takes too much memory. Training is too slow.

Bootstrapping 1.(Initialization:) Choose some training examples. Not too few, not too many. 2.Train a detector. 3.Apply the detector to all training images. 4.Identify mistakes. 5.Add mistakes to training examples. If needed, remove some examples to make room. 6.Go to step 2, unless performance has stopped improving.

Bootstrapping Discussion Allows gradual construction of a good training set. –It is important to include difficult examples in training. Keeps overall memory and time requirements in check. You should stop when performance stops improving. –Performance can be measured on the entire training set.

Classifier Cascades A generalization of the filter-and-refine approach. Goal: improve detection efficiency, without sacrificing (too much) accuracy. A cascade is a sequence of classifiers –C 1, C 2, …, C K. Each C i+1 is slower and more accurate than C i.

Applying a Cascade to a Window function result = cascade_classify(window) For i = 1 to K-1. –If C i (window) is less than a threshold T i : Reject window (in other words, return “not a face”) and quit. Return C K (window).

Cascade Discussion function result = cascade_classify(window) For i = 1 to K-1. –If C i (window) is less than a threshold T i : Reject window (in other words, return “not a face”) and quit. Return C K (window). A cascade returns “face” if: –None of the first K-1 classifiers rejects the window, AND –C K classifies the window as a face. A cascade returns “not a face” if: –Any of the first K-1 classifiers rejects the window, OR –C K classifies the window as a non-face.

Why Are Cascades Faster? function result = cascade_classify(window) For i = 1 to K-1. –If C i (window) is less than a threshold T i : Reject window (in other words, return “not a face”) and quit. Return C K (window). Consider C 1. –It is applied to ALL image windows. –For most windows, it is easy to tell that they are NOT faces. –C 1 can be relatively simple. –Key requirement: MINIMIZE REJECTION OF FACES. –Threshold T 1 can be tuned to achieve the requirement.

Example Suppose that our strong classifier H contains 30 weak classifiers. Construct C 1 using only the first three weak classifiers. C 1 can be applied pretty fast to each window. Suppose that we find a T 1 so that no faces are rejected, and 70% of non-faces are rejected. How much faster is it to use a cascade of C 1 and H instead of using just H?

Example Suppose that our strong classifier H contains 30 weak classifiers. Construct C 1 using only the first three weak classifiers. C 1 can be applied pretty fast to each window. Suppose that we find a T 1 so that no faces are rejected, and 70% of non-faces are rejected. How much faster is it to use a cascade of C 1 and H instead of using just H? –Cascade running time: 100% * 3/ % * 30/30 = 40% of running time of H. Adding more classifiers can make it even faster.

Constructing a Cascade If strong classifier consists of 30 weak classifiers: –C 1 can be the sum of the first 3 weak classifiers. –C 2 can be the sum of the first 7 weak classifiers. –C 3 can be the sum of the first 10 weak classifiers. Important: in computing C i, REUSE the results of C i-1, that have already been computed. How do we choose numbers 3, 7, 10? How do we choose how many classifiers to include? –By trial and error.

Constructing a Cascade If strong classifier consists of 30 weak classifiers: –C 1 can be the sum of the first 3 weak classifiers. –C 2 can be the sum of the first 7 weak classifiers. –C 3 can be the sum of the first 10 weak classifiers. How do we choose thresholds T i ? –Each classifier gives a score for each image. More positive scores imply that the image looks more like a face. –Find a T i such that almost no face examples score below T i. What does “almost no face examples” mean? –That is up to the designer of the system. As an example, it can be interpreted as “not more than 0.1%”.