Presentation is loading. Please wait.

Presentation is loading. Please wait.

Oisin Mac Aodha (UCL) Gabriel Brostow (UCL) Marc Pollefeys (ETH)

Similar presentations


Presentation on theme: "Oisin Mac Aodha (UCL) Gabriel Brostow (UCL) Marc Pollefeys (ETH)"— Presentation transcript:

1 Oisin Mac Aodha (UCL) Gabriel Brostow (UCL) Marc Pollefeys (ETH)

2 Video from Dorothy Kuipers

3  #2 all-time Computer Vision problem (disputable)  “Where did each pixel go?”

4  Compared against each other on the “blind” Middlebury test set

5 Anonymous. Secrets of optical flow estimation and their principles. CVPR 2010 submission 477

6

7

8

9 Should use algorithm A Should use algorithm B

10

11 (Artistic version; object-boundaries don’t interest us)

12  that the most suitable algorithm can be chosen for each video automatically, through supervised training of a classifier

13  that one can predict the space-time segments of the video that are best-served by each available algorithm  (Can always come back to choose a per-frame or per-video algorithm)

14 Image data Groundtruth labels Learning algorithm Feature extraction Feature extraction Trained classifier Trained classifier Estimated class labels Estimated class labels

15 Image data Groundtruth labels Learning algorithm Feature extraction Feature extraction Estimated class labels Estimated class labels Trained classifier Trained classifier New test data New test data

16 Image data Groundtruth labels Learning algorithm Feature extraction Feature extraction Estimated class labels Estimated class labels Trained classifier Trained classifier New test data New test data Random Forests Breiman, 2001

17 Image data Groundtruth labels Learning algorithm Feature extraction Feature extraction Estimated class labels Estimated class labels Trained classifier Trained classifier New test data New test data

18

19

20  Training data D consists of feature vectors x and class labels c (i.e. best-algorithm per pixel)  Feature vector x is multi-scale, and includes:  Spatial Gradient  Distance Transform  Temporal Gradient  Residual Error (after bicubic reconstruction)

21  Training data D consists of feature vectors x and class labels c (i.e. best-algorithm per pixel)  Feature vector x is multi-scale, and includes:  Spatial Gradient  Distance Transform  Temporal Gradient  Residual Error (after bicubic reconstruction)

22  Training data D consists of feature vectors x and class labels c (i.e. best-algorithm per pixel)  Feature vector x is multi-scale, and includes:  Spatial Gradient  Distance Transform  Temporal Gradient  Residual Error (after bicubic reconstruction)

23

24  Temporal Gradient  Residual Error

25

26

27 False Positive Rate (for EPE < 1.0) FlowLib Decision Confidence True Po- sitive Rate Ave EPE Number of pixels x10 5

28

29  What is a match? Details are important…  Nearest neighbor (see also FLANN)FLANN  Distance Ratio  PCA  Evaluation: density, # correct matches, tolerance “192 correct matches (yellow) and 208 false matches (blue)”

30

31

32

33  Current results don’t quite live up to the theory:  Flaws of best-algorithm are the upper bound (ok)  Training data does not fit in memory (fixable)  “Winning” the race is more than rank (problem!)

34  Overall, predictions are correlated with the best algorithm for each segment (expressed as Pr!)  Training data where one class dominates is dangerous – needs improvement  Other features could help make better predictions  Results don’t yet do the idea justice  One size does NOT fit all  At least in terms of algorithm suitability  Could use “bad” algorithms!

35

36

37 Ground Truth BestBased on Prediction

38 White = 30 pixel end point error FlowLibBased on Prediction

39 FlowLibBased on Prediction (Contrast enhanced)


Download ppt "Oisin Mac Aodha (UCL) Gabriel Brostow (UCL) Marc Pollefeys (ETH)"

Similar presentations


Ads by Google