Presentation is loading. Please wait.

Presentation is loading. Please wait.

Ensemble Tracking Shai Avidan IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE February 2007.

Similar presentations


Presentation on theme: "Ensemble Tracking Shai Avidan IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE February 2007."— Presentation transcript:

1 Ensemble Tracking Shai Avidan IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE February 2007

2 outline Prior knowledge : Adaboost Introduction Ensemble tracking Implementation issues Experiments

3 Adaboost Resampling for Classifier Design  Bagging Use multiple versions of a training set Each created by drawing n ’ <n samples from D with replacement (i.e. if a sample is drawn, it is not removed from D but is reconsidered in the next sampling) Each data set is used to train a different component classifier The final classification decision is based on the vote of the component classifiers

4 Adaboost  Boosting To generate complementary classifiers by training the next component classifier on the mistakes of the previous ones Using a subset of the training data that is most informative given the current set of component classifiers Adaboost trains a weak classifier on increasingly more difficult examples and combines the result to produce a strong classifier that is better than any of the weak classifiers. Weak classifier : Strong classifier :

5 Adaboost AdaBoost(adaptive boosting)  Use the same training set over and over  Each training pattern receives a weight W k (i) The probability that the i-th pattern is drawn to take the kth component classifier. Uniform initialization W 1 (i)=1/n If a training pattern is accurately classified h k (x i )=y i, its chance of used again is reduced Otherwise, h k (x i )  y i

6 Adaboost  Final decision

7 Adaboost K max component classifiers

8 Adaboost At the t step

9 Adaboost

10

11 Introduction Considering tracking as a binary classification problem. Ensemble tracking as a method for training classifiers on time-varying distributions. Ensemble of weak classifiers is trained online to distinguish between the object and the background.

12 Introduction

13

14 Ensemble tracking maintains an implicit representation of the foreground and the background instead of describing foreground object explicitly alone. Ensemble is not template-based methods. Those maintains the spatial integrity of the objects and are especially suited for handling rigid objects.

15 Introduction Ensemble tracking extends traditional mean-shift tracking in a number of important directions:  Mean-shift tracking usually works with histogram of RGB colors. This is because gray- scale images do not provide enough information for tracking and high-dimensional feature spaces cannot be modeled with histograms due to exponential memory requirements.

16 Introduction This is in contrast to existing methods that either represent the foreground object using the most recent histogram or some ad hoc combination of the histograms of the first and last frames.

17 Introduction Other advantages:  It breaks the time consuming training phase into a sequence of simple and easy to compute learning tasks that can be performed online.  It can also integrate offline and online learning seamlessly.  Integrating classifier over time improves the stability of the tracker in cases of partial occlusions or illumination changes.

18 In each frame, we keep the K “ best ” weak classifiers, discard the remaining T-K new weak classifiers, train T-K new weak classifiers on the newly available data, and reconstruct the strong weak classifier. The margin of the weak classifier h(x) is mapped to a confidence measure c(x) by clipping negative margins to zero and rescaling the positive margins to the range [0,1].

19 Ensemble update

20 Ensemble tracking

21

22 During Step 7 of choosing the K best weak classifier, weak classifiers do not perform much better than chance. We allow up to existing weak classifiers to be removed this way because a large number might be a sign of occlusion and keep the ensemble unchanged for this frame.

23 Implementations issues Outlier Rejection

24 Implementations issues

25 Multiresolution Tracking

26 Implementations issues

27 experiments The first version uses five weak classifiers, each working on an 11D feature vector per pixel that consists of an 8-bin local histogram of oriented gradients calculated on a 5x5 window as well as the pixel R, G, and B valuse. To improve robustness, we only count edges that are above some predefined threshold, which war set to 10 intensity values.

28 experiments We found that the original feature space was not stable enough and used a nonlinear version of that feature space instead. We use only three, instead of five weak classifiers. Three levels of the pyramid In each frame, we drop one weak classifier and add a newly trained weak classifier.

29 experiments We allow the tracker to drop up to two weak classifiers per frame because dropping more than that might be could be a sign of occlusion and we therefore do not update the ensemble in such a case.

30 experiments Results on Color Sequences: a pedestrian crossing the streat

31 experiments Results on Color Sequences: tracking a couple walking with a hand-held camera.

32 experiments Results on Color Sequences: tracking a face exhibiting out-of-plane rotations

33 experiments Results on Color Sequences: tracking a red car that is undergoing out-of- plane rotations and partial occlusions. 11D feature vector, single scale, an ensemble of three classifier was enough to obtain robust and stable tracking

34 experiments Analyze the importance of the update scheme for tracking:

35 experiments Analyze how often are the weak classifiers updated?

36 experiments Analyze how does their weight change over time.

37 experiments Analyze how does this method compare with a standard AdaBoost classifier that trains all its weak classifiers on a given frame?

38 experiments Results on gray-scale sequence :

39 experiments Results on IR sequence:

40 experiments Handling long-period occlusion  Classification rate is the fraction of the number pixels that were correctly classified  As long as the classification rate is high, the tracking goes unchanged.  When the classification level drops(<0.5), switch to prediction mode.  Once occlusion is detected we start sampling, according to the particle filter, possible location where the object might appear.  In each such location, compute the classification score. If it is above a threshold (0.7), then tracking resumes.

41 experiments Handling occlusions:

42 experiments


Download ppt "Ensemble Tracking Shai Avidan IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE February 2007."

Similar presentations


Ads by Google