Ensemble Tracking Shai Avidan IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE February 2007.

Slides:



Advertisements
Similar presentations
Ensemble Learning Reading: R. Schapire, A brief introduction to boosting.
Advertisements

Change Detection C. Stauffer and W.E.L. Grimson, “Learning patterns of activity using real time tracking,” IEEE Trans. On PAMI, 22(8): , Aug 2000.
Computer Vision for Human-Computer InteractionResearch Group, Universität Karlsruhe (TH) cv:hci Dr. Edgar Seemann 1 Computer Vision: Histograms of Oriented.
Presenter: Hoang, Van Dung
AdaBoost & Its Applications
Longin Jan Latecki Temple University
Foreground Modeling The Shape of Things that Came Nathan Jacobs Advisor: Robert Pless Computer Science Washington University in St. Louis.
A Nonparametric Treatment for Location/Segmentation Based Visual Tracking Le Lu Integrated Data Systems Dept. Siemens Corporate Research, Inc. Greg Hager.
Face detection Many slides adapted from P. Viola.
EE462 MLCV Lecture 5-6 Object Detection – Boosting Tae-Kyun Kim.
Robust Moving Object Detection & Categorization using self- improving classifiers Omar Javed, Saad Ali & Mubarak Shah.
Modeling Pixel Process with Scale Invariant Local Patterns for Background Subtraction in Complex Scenes (CVPR’10) Shengcai Liao, Guoying Zhao, Vili Kellokumpu,
Motion Detection And Analysis Michael Knowles Tuesday 13 th January 2004.
Recovering Intrinsic Images from a Single Image 28/12/05 Dagan Aviv Shadows Removal Seminar.
Generic Object Detection using Feature Maps Oscar Danielsson Stefan Carlsson
A Brief Introduction to Adaboost
Ensemble Learning: An Introduction
Adaboost and its application
A Robust Real Time Face Detection. Outline  AdaBoost – Learning Algorithm  Face Detection in real life  Using AdaBoost for Face Detection  Improvements.
Introduction to Boosting Aristotelis Tsirigos SCLT seminar - NYU Computer Science.
Smart Traveller with Visual Translator for OCR and Face Recognition LYU0203 FYP.
Face Detection using the Viola-Jones Method
Prakash Chockalingam Clemson University Non-Rigid Multi-Modal Object Tracking Using Gaussian Mixture Models Committee Members Dr Stan Birchfield (chair)
Olga Zoidi, Anastasios Tefas, Member, IEEE Ioannis Pitas, Fellow, IEEE
BraMBLe: The Bayesian Multiple-BLob Tracker By Michael Isard and John MacCormick Presented by Kristin Branson CSE 252C, Fall 2003.
Introduction to Computer Vision Olac Fuentes Computer Science Department University of Texas at El Paso El Paso, TX, U.S.A.
1. Introduction Motion Segmentation The Affine Motion Model Contour Extraction & Shape Estimation Recursive Shape Estimation & Motion Estimation Occlusion.
A General Framework for Tracking Multiple People from a Moving Camera
Detecting Pedestrians Using Patterns of Motion and Appearance Paul Viola Microsoft Research Irfan Ullah Dept. of Info. and Comm. Engr. Myongji University.
Window-based models for generic object detection Mei-Chen Yeh 04/24/2012.
Object Recognition in Images Slides originally created by Bernd Heisele.
ECE738 Advanced Image Processing Face Detection IEEE Trans. PAMI, July 1997.
Ensemble Learning Spring 2009 Ben-Gurion University of the Negev.
CSCE 643 Computer Vision: Extractions of Image Features Jinxiang Chai.
21 June 2009Robust Feature Matching in 2.3μs1 Simon Taylor Edward Rosten Tom Drummond University of Cambridge.
Limitations of Cotemporary Classification Algorithms Major limitations of classification algorithms like Adaboost, SVMs, or Naïve Bayes include, Requirement.
Tony Jebara, Columbia University Advanced Machine Learning & Perception Instructor: Tony Jebara.
Face Detection Ying Wu Electrical and Computer Engineering Northwestern University, Evanston, IL
Face Detection Using Large Margin Classifiers Ming-Hsuan Yang Dan Roth Narendra Ahuja Presented by Kiang “Sean” Zhou Beckman Institute University of Illinois.
Expectation-Maximization (EM) Case Studies
Histograms of Oriented Gradients for Human Detection(HOG)
E NSEMBLE L EARNING : A DA B OOST Jianping Fan Dept of Computer Science UNC-Charlotte.
The Viola/Jones Face Detector A “paradigmatic” method for real-time object detection Training is slow, but detection is very fast Key ideas Integral images.
Bibek Jang Karki. Outline Integral Image Representation of image in summation format AdaBoost Ranking of features Combining best features to form strong.
Boosted Particle Filter: Multitarget Detection and Tracking Fayin Li.
1 CHUKWUEMEKA DURUAMAKU.  Machine learning, a branch of artificial intelligence, concerns the construction and study of systems that can learn from data.
Ensemble Methods in Machine Learning
Course14 Dynamic Vision. Biological vision can cope with changing world Moving and changing objects Change illumination Change View-point.
FACE DETECTION : AMIT BHAMARE. WHAT IS FACE DETECTION ? Face detection is computer based technology which detect the face in digital image. Trivial task.
Week 10 Emily Hand UNR.
1 January 24, 2016Data Mining: Concepts and Techniques 1 Data Mining: Concepts and Techniques — Chapter 7 — Classification Ensemble Learning.
Project Overview CSE 6367 – Computer Vision Vassilis Athitsos University of Texas at Arlington.
Face Detection and Head Tracking Ying Wu Electrical Engineering & Computer Science Northwestern University, Evanston, IL
Max-Confidence Boosting With Uncertainty for Visual tracking WEN GUO, LIANGLIANG CAO, TONY X. HAN, SHUICHENG YAN AND CHANGSHENG XU IEEE TRANSACTIONS ON.
AdaBoost Algorithm and its Application on Object Detection Fayin Li.
Learning and Removing Cast Shadows through a Multidistribution Approach Nicolas Martel-Brisson, Andre Zaccarin IEEE TRANSACTIONS ON PATTERN ANALYSIS AND.
Detecting Occlusion from Color Information to Improve Visual Tracking
Motion tracking TEAM D, Project 11: Laura Gui - Timisoara Calin Garboni - Timisoara Peter Horvath - Szeged Peter Kovacs - Debrecen.
Another Example: Circle Detection
Machine Learning: Ensemble Methods
Reading: R. Schapire, A brief introduction to boosting
Cascade for Fast Detection
Motion Detection And Analysis
Lit part of blue dress and shadowed part of white dress are the same color
COMP61011 : Machine Learning Ensemble Models
Feature description and matching
Data Mining Practical Machine Learning Tools and Techniques
Cos 429: Face Detection (Part 2) Viola-Jones and AdaBoost Guest Instructor: Andras Ferencz (Your Regular Instructor: Fei-Fei Li) Thanks to Fei-Fei.
ADABOOST(Adaptative Boosting)
Feature descriptors and matching
Presentation transcript:

Ensemble Tracking Shai Avidan IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE February 2007

outline Prior knowledge : Adaboost Introduction Ensemble tracking Implementation issues Experiments

Adaboost Resampling for Classifier Design  Bagging Use multiple versions of a training set Each created by drawing n ’ <n samples from D with replacement (i.e. if a sample is drawn, it is not removed from D but is reconsidered in the next sampling) Each data set is used to train a different component classifier The final classification decision is based on the vote of the component classifiers

Adaboost  Boosting To generate complementary classifiers by training the next component classifier on the mistakes of the previous ones Using a subset of the training data that is most informative given the current set of component classifiers Adaboost trains a weak classifier on increasingly more difficult examples and combines the result to produce a strong classifier that is better than any of the weak classifiers. Weak classifier : Strong classifier :

Adaboost AdaBoost(adaptive boosting)  Use the same training set over and over  Each training pattern receives a weight W k (i) The probability that the i-th pattern is drawn to take the kth component classifier. Uniform initialization W 1 (i)=1/n If a training pattern is accurately classified h k (x i )=y i, its chance of used again is reduced Otherwise, h k (x i )  y i

Adaboost  Final decision

Adaboost K max component classifiers

Adaboost At the t step

Adaboost

Introduction Considering tracking as a binary classification problem. Ensemble tracking as a method for training classifiers on time-varying distributions. Ensemble of weak classifiers is trained online to distinguish between the object and the background.

Introduction

Ensemble tracking maintains an implicit representation of the foreground and the background instead of describing foreground object explicitly alone. Ensemble is not template-based methods. Those maintains the spatial integrity of the objects and are especially suited for handling rigid objects.

Introduction Ensemble tracking extends traditional mean-shift tracking in a number of important directions:  Mean-shift tracking usually works with histogram of RGB colors. This is because gray- scale images do not provide enough information for tracking and high-dimensional feature spaces cannot be modeled with histograms due to exponential memory requirements.

Introduction This is in contrast to existing methods that either represent the foreground object using the most recent histogram or some ad hoc combination of the histograms of the first and last frames.

Introduction Other advantages:  It breaks the time consuming training phase into a sequence of simple and easy to compute learning tasks that can be performed online.  It can also integrate offline and online learning seamlessly.  Integrating classifier over time improves the stability of the tracker in cases of partial occlusions or illumination changes.

In each frame, we keep the K “ best ” weak classifiers, discard the remaining T-K new weak classifiers, train T-K new weak classifiers on the newly available data, and reconstruct the strong weak classifier. The margin of the weak classifier h(x) is mapped to a confidence measure c(x) by clipping negative margins to zero and rescaling the positive margins to the range [0,1].

Ensemble update

Ensemble tracking

During Step 7 of choosing the K best weak classifier, weak classifiers do not perform much better than chance. We allow up to existing weak classifiers to be removed this way because a large number might be a sign of occlusion and keep the ensemble unchanged for this frame.

Implementations issues Outlier Rejection

Implementations issues

Multiresolution Tracking

Implementations issues

experiments The first version uses five weak classifiers, each working on an 11D feature vector per pixel that consists of an 8-bin local histogram of oriented gradients calculated on a 5x5 window as well as the pixel R, G, and B valuse. To improve robustness, we only count edges that are above some predefined threshold, which war set to 10 intensity values.

experiments We found that the original feature space was not stable enough and used a nonlinear version of that feature space instead. We use only three, instead of five weak classifiers. Three levels of the pyramid In each frame, we drop one weak classifier and add a newly trained weak classifier.

experiments We allow the tracker to drop up to two weak classifiers per frame because dropping more than that might be could be a sign of occlusion and we therefore do not update the ensemble in such a case.

experiments Results on Color Sequences: a pedestrian crossing the streat

experiments Results on Color Sequences: tracking a couple walking with a hand-held camera.

experiments Results on Color Sequences: tracking a face exhibiting out-of-plane rotations

experiments Results on Color Sequences: tracking a red car that is undergoing out-of- plane rotations and partial occlusions. 11D feature vector, single scale, an ensemble of three classifier was enough to obtain robust and stable tracking

experiments Analyze the importance of the update scheme for tracking:

experiments Analyze how often are the weak classifiers updated?

experiments Analyze how does their weight change over time.

experiments Analyze how does this method compare with a standard AdaBoost classifier that trains all its weak classifiers on a given frame?

experiments Results on gray-scale sequence :

experiments Results on IR sequence:

experiments Handling long-period occlusion  Classification rate is the fraction of the number pixels that were correctly classified  As long as the classification rate is high, the tracking goes unchanged.  When the classification level drops(<0.5), switch to prediction mode.  Once occlusion is detected we start sampling, according to the particle filter, possible location where the object might appear.  In each such location, compute the classification score. If it is above a threshold (0.7), then tracking resumes.

experiments Handling occlusions:

experiments