Real-Time Object Localization and Tracking from Image Sequences

Slides:



Advertisements
Similar presentations
Bayesian Belief Propagation
Advertisements

Database-Based Hand Pose Estimation CSE 6367 – Computer Vision Vassilis Athitsos University of Texas at Arlington.
Forward-Backward Correlation for Template-Based Tracking Xiao Wang ECE Dept. Clemson University.
Robust Object Tracking via Sparsity-based Collaborative Model
Adviser : Ming-Yuan Shieh Student ID : M Student : Chung-Chieh Lien VIDEO OBJECT SEGMENTATION AND ITS SALIENT MOTION DETECTION USING ADAPTIVE BACKGROUND.
Optimization & Learning for Registration of Moving Dynamic Textures Junzhou Huang 1, Xiaolei Huang 2, Dimitris Metaxas 1 Rutgers University 1, Lehigh University.
Formation et Analyse d’Images Session 8
Modeling Pixel Process with Scale Invariant Local Patterns for Background Subtraction in Complex Scenes (CVPR’10) Shengcai Liao, Guoying Zhao, Vili Kellokumpu,
Motion Tracking. Image Processing and Computer Vision: 82 Introduction Finding how objects have moved in an image sequence Movement in space Movement.
Motion Detection And Analysis Michael Knowles Tuesday 13 th January 2004.
Adam Rachmielowski 615 Project: Real-time monocular vision-based SLAM.
Introduction to Kalman Filter and SLAM Ting-Wei Hsu 08/10/30.
Probabilistic video stabilization using Kalman filtering and mosaicking.
Adaptive Rao-Blackwellized Particle Filter and It’s Evaluation for Tracking in Surveillance Xinyu Xu and Baoxin Li, Senior Member, IEEE.
1 Integration of Background Modeling and Object Tracking Yu-Ting Chen, Chu-Song Chen, Yi-Ping Hung IEEE ICME, 2006.
Student: Hsu-Yung Cheng Advisor: Jenq-Neng Hwang, Professor
Jacinto C. Nascimento, Member, IEEE, and Jorge S. Marques
1 Video Surveillance systems for Traffic Monitoring Simeon Indupalli.
Overview and Mathematics Bjoern Griesbach
Adaptive Signal Processing Class Project Adaptive Interacting Multiple Model Technique for Tracking Maneuvering Targets Viji Paul, Sahay Shishir Brijendra,
1 Formation et Analyse d’Images Session 7 Daniela Hall 7 November 2005.
Kalman filter and SLAM problem
Tracking Pedestrians Using Local Spatio- Temporal Motion Patterns in Extremely Crowded Scenes Louis Kratz and Ko Nishino IEEE TRANSACTIONS ON PATTERN ANALYSIS.
3D Fingertip and Palm Tracking in Depth Image Sequences
Prakash Chockalingam Clemson University Non-Rigid Multi-Modal Object Tracking Using Gaussian Mixture Models Committee Members Dr Stan Birchfield (chair)
1. Introduction Motion Segmentation The Affine Motion Model Contour Extraction & Shape Estimation Recursive Shape Estimation & Motion Estimation Occlusion.
A General Framework for Tracking Multiple People from a Moving Camera
Optical Flow Donald Tanguay June 12, Outline Description of optical flow General techniques Specific methods –Horn and Schunck (regularization)
3D SLAM for Omni-directional Camera
Video Tracking Using Learned Hierarchical Features
Dynamic 3D Scene Analysis from a Moving Vehicle Young Ki Baik (CV Lab.) (Wed)
Complete Pose Determination for Low Altitude Unmanned Aerial Vehicle Using Stereo Vision Luke K. Wang, Shan-Chih Hsieh, Eden C.-W. Hsueh 1 Fei-Bin Hsaio.
Young Ki Baik, Computer Vision Lab.
Forward-Scan Sonar Tomographic Reconstruction PHD Filter Multiple Target Tracking Bayesian Multiple Target Tracking in Forward Scan Sonar.
A Fast and Accurate Tracking Algorithm of the Left Ventricle in 3D Echocardiography A Fast and Accurate Tracking Algorithm of the Left Ventricle in 3D.
December 9, 2014Computer Vision Lecture 23: Motion Analysis 1 Now we will talk about… Motion Analysis.
1 University of Texas at Austin Machine Learning Group 图像与视频处理 计算机学院 Motion Detection and Estimation.
Stable Multi-Target Tracking in Real-Time Surveillance Video
1 Research Question  Can a vision-based mobile robot  with limited computation and memory,  and rapidly varying camera positions,  operate autonomously.
Expectation-Maximization (EM) Case Studies
Chapter 5 Multi-Cue 3D Model- Based Object Tracking Geoffrey Taylor Lindsay Kleeman Intelligent Robotics Research Centre (IRRC) Department of Electrical.
Spatiotemporal Saliency Map of a Video Sequence in FPGA hardware David Boland Acknowledgements: Professor Peter Cheung Mr Yang Liu.
Implicit Active Shape Models for 3D Segmentation in MR Imaging M. Rousson 1, N. Paragio s 2, R. Deriche 1 1 Odyssée Lab., INRIA Sophia Antipolis, France.
Tracking Groups of People for Video Surveillance Xinzhen(Elaine) Wang Advisor: Dr.Longin Latecki.
Representing Moving Images with Layers J. Y. Wang and E. H. Adelson MIT Media Lab.
Minimum Barrier Salient Object Detection at 80 FPS JIANMING ZHANG, STAN SCLAROFF, ZHE LIN, XIAOHUI SHEN, BRIAN PRICE, RADOMIR MECH IEEE INTERNATIONAL CONFERENCE.
Kalman Filter and Data Streaming Presented By :- Ankur Jain Department of Computer Science 7/21/03.
Zhaoxia Fu, Yan Han Measurement Volume 45, Issue 4, May 2012, Pages 650–655 Reporter: Jing-Siang, Chen.
Detecting Occlusion from Color Information to Improve Visual Tracking
Video object segmentation and its salient motion detection using adaptive background generation Kim, T.K.; Im, J.H.; Paik, J.K.;  Electronics Letters 
V4 – Video Tracker for Extremely Hard Night Conditions
Signal and Image Processing Lab
Paper – Stephen Se, David Lowe, Jim Little
Tracking Objects with Dynamics
Motion Detection And Analysis
Fast and Robust Object Tracking with Adaptive Detection
Enhanced-alignment Measure for Binary Foreground Map Evaluation
ISOMAP TRACKING WITH PARTICLE FILTERING
Representing Moving Images with Layers
Representing Moving Images with Layers
Presented by: Yang Yu Spatiotemporal GMM for Background Subtraction with Superpixel Hierarchy Mingliang Chen, Xing Wei, Qingxiong.
Combining Geometric- and View-Based Approaches for Articulated Pose Estimation David Demirdjian MIT Computer Science and Artificial Intelligence Laboratory.
PRAKASH CHOCKALINGAM, NALIN PRADEEP, AND STAN BIRCHFIELD
Image and Video Processing
Introduction to Object Tracking
Optical flow and keypoint tracking
Tracking Many slides adapted from Kristen Grauman, Deva Ramanan.
Multi-UAV to UAV Tracking
Report 2 Brandon Silva.
Presentation transcript:

Real-Time Object Localization and Tracking from Image Sequences Yuanwei Wu, Yao Sui, Arjan Gupta and Guanghui Wang Friday, Sep. 9, 2016

Background Amazon: On December 1, 2013, the CEO of Amazon Jeff Bezos revealed their plans about future delivery system using small unmanned aerial vehicles (UAVs) technology. Real-time and Autonomous sense and avoidance navigation system Vision-based methods: Robust to electromagnetic interference Compact and low power consumption Source: Amazon, “Amazon prime air,” https://www.youtube.com/watch?v=98BIu9dpwHU, 2013. Robust to electromagnetic interference compared to conventional sensor-based methods

Salient Object Detection and Tracking The task of salient object detection is to compute a saliency map and segment an accurate boundary of that object. This amazon UAV sequence with challenging situations, namely scale variation, out-of-view and re-appearance. For detection, MB+ fails to provide a high quality saliency map.

Salient Object Detection and Tracking The goal of visual tracking is to estimate the boundary and trajectory of the object in every frame of an image sequence. For tracking, the existing trackers cannot handle out-of-view and re-appearance challenges. Our method provides high quality of saliency map in detection, and accurate scale and position of the target in tracking.

Previous Works Real-time Automatic Initialization Detection and tracking [Andriluka et al. CVPR’08]  Saliency-based tracking [Mahadevan et al. CVPR’09]  [Andriluka et al. CVPR’08]: combine a detector with a tracker, however, it needs large amount of off-line training for pedestrians. [Mahadevan et al. CVPR’09]: utilize the center surround contrast cue to calculate the saliency map and discriminate object from background, however it builds the motion saliency maps using optical flow makes it computational intensive.

Previous Works Real-time Automatic Initialization Detection and tracking [Andriluka et al. CVPR’08]  Saliency-based tracking [Mahadevan et al. CVPR’09]  State-of-the-art Tracking-by-detection CT [Zhang et al. ECCV’12] STC [Zhang et al. ECCV’14] CN[Danelljan et al. TPAMI’14] SAMF[Li et al. ECCVW’14] DSST[Danelljan et al. BMVC’14] CCT[Zhu et al. BMVC’15] KCF[Henriques et al. TPAMI’15] Real time trackers: CT: sparsity-based compressive tracking, no scale adaptive bounding box Correlation filter based trackers: STC, CN, SAMF, DSST, CCT, KCF

Previous Works Real-time Automatic Initialization Detection and tracking [Andriluka et al. CVPR’08]  Saliency-based tracking [Mahadevan et al. CVPR’09]  State-of-the-art Tracking-by-detection CT [Zhang et al. ECCV’12] STC [Zhang et al. ECCV’14] CN[Danelljan et al. TPAMI’14] SAMF[Li et al. ECCVW’14] DSST[Danelljan et al. BMVC’14] CCT[Zhu et al. BMVC’15] KCF[Henriques et al. TPAMI’15] Detection-then-tracking Proposed Proposed approach: real-time and automatic initialization by integration of kalman filter and salient object detection

Contributions The proposed algorithm integrates saliency map into a dynamic model and adopts the target-specific saliency map as the observation for tracking; Developed a tracker with automatic initialization for real-world applications; The proposed technique achieves state-of-the-art performance from extensive real experiments. Contributions:

System Overview Fast Object Localization and Tracking (FOLT) In this approach, the bounding box of the object is initialized from the saliency map of the entire image. A dynamic motion is established to predict the object position and size at the next frame. After initialization, the proposed approach runs recursively on prediction, observation and correction phases.

System Overview: prediction Coarse solution: linear Gaussian motion model [24] motion state: , where (x, y) denotes the center coordinates, (u, v) denotes the velocities, and (w, h) denotes the width and height of the minimum bounding box. Project the state ahead: (1), where is additive white Gaussian noise. Project the error covariance ahead: In most tracking scenarios, a linear Gaussian motion model has been demonstrated to be an eective representation for the motion behavior of a salient object in natural image sequences [24, 36]. Under the constraint of natural motion, this predicted bounding box provides the tracking algorithm a coarse solution, which is not far away from the ground truth[24]. [24] Yin, S., Na, J.H., Choi, J.Y., Oh, S.: "hierarchical kalman-particle flter with adaptation to motion changes for object tracking“, CVIU 115(6) (2011) pp. 885-900

System Overview: observation Refine solution Observation state: A search region is automatically attained by expanding the predicted bounding box with a fixed percentage. The location and size of the object is refined by computing the saliency within the search region, and thresholding the saliency map. The observation zt is the output of the fast salient object detector.

System Overview: correction Update the motion model Compute the Kalman gain Update estimate with measurement (2) Update the error covariance (a) Prediction (a) Observation (c) Correction Next, the refined bounding box, as a new observation, is fed to the Kalman filter to update the motion model in the correction phase.

Salient Object Detection Measuring Boundary Connectivity by Distance Transform Compute the distance for each pixel w.r.t. the image boundary Seed Set Shortest Path The idea of MBD: How to measure image boundary connectivity by distance transform: Set image boundary pixels as seed set (show in red) For each pixel (show in greed), find the shortest path (show in grey) to the seed set, according to the given path cost function The cost of the shortest path is the distance between green and red. Source: Zhang et al. ICCV’15

MBD vs Geodesic Distance In what follows, we consider a single-channel image. MBD [Strand et al. CVIU’13] Geodesic + = Introducing MBD We consider a single-channel real-valued image The path cost function of MBD: \pi is the path, ie a sequence of adjacent pixels I(\pi(i)) represents the pixel value for the i-th pixel on the path The path cost function measure the distance between the highest and the lowest point along the path. The path cost function of Geodesic Distance Source: Zhang et al. ICCV’15

MBD is robust to small pixel value fluctuation MBD vs Geodesic Distance MBD is robust to small pixel value fluctuation MBD MBD is more robust to small pixel fluctuation: MBD With higher sampling frequency (higher resolution), pixel values a long the path can have more fluctuations. The path cost function of MBD is robust to this, because it only depends on the lowest and the highest points of the path Geodesic However, the path cost function is sensitive to small pixel value fluctuation, as they will accumulate along the path. When applied on raw image pixels, the effectiveness of the geodesic distance can be greatly affected by this phenomenon. = Geodesic + = Source: Zhang et al. ICCV’15

Raster scanning/Inverse-raster scanning For tracking, the existing trackers cannot handle out-of-view and re-appearance challenges. Our method provides high quality of saliency map in detection, and accurate scale and position of the target in tracking.

Algorithm y x For each visited pixel x: Check each of the 4-connected neighbors x y Algorithm: For each visited pixel (green), check each of its 4-connected neighbor during forward and backward pass Source: Zhang et al. ICCV’15

Algorithm For each visited pixel x: Check each of the 4-connected neighbors Minimize the path cost Update: D(x), cost of current assigned path U(x), highest value on assigned path L(x), lowest value on assigned path D(y) U(y) L(y) D(x) U(x) L(x) During the passes, we keep track of the following relevant information for each pixel: The path cost of the currently assigned path: D(x) The highest value on the assigned path: U(x) The lowest value on the assigned path: L(x) Then we can minimize the path cost by checking the neighbors using the displayed formulation After the minimization, we upate D(x), L(x) and U(x) Source: Zhang et al. ICCV’15

Algorithm Combined For tracking, the existing trackers cannot handle out-of-view and re-appearance challenges. Our method provides high quality of saliency map in detection, and accurate scale and position of the target in tracking.

Experiments Parameters: Transition state matrix Measure matrix For tracking, the existing trackers cannot handle out-of-view and re-appearance challenges. Our method provides high quality of saliency map in detection, and accurate scale and position of the target in tracking.

Experiments Parameters: Process noise covariance matrix Measure noise covariance matrix For tracking, the existing trackers cannot handle out-of-view and re-appearance challenges. Our method provides high quality of saliency map in detection, and accurate scale and position of the target in tracking. Error covariance pre-state

Qualitative Evaluation For tracking, the existing trackers cannot handle out-of-view and re-appearance challenges. Our method provides high quality of saliency map in detection, and accurate scale and position of the target in tracking.

Precision and Success Plots One pass evaluation: Temporal robustness evaluation: TRE For tracking, the existing trackers cannot handle out-of-view and re-appearance challenges. Our method provides high quality of saliency map in detection, and accurate scale and position of the target in tracking.

Precision and Success Rates on 15 Sequences

Precision and Success Rates on 15 Sequences

Qualitative Evaluation Illumination variation In-plane and out-of-plane rotations Scale variation

Limitations

Conclusions In this paper, we have proposed an effective and efficient approach for real-time visual object localization and tracking Our method integrates a fast salient object detector within Kalman filtering framework. Compared to the state-of-the-art trackers, our approach can not only initialize automatically, it also achieves the fastest speed and better performance than competing trackers. For tracking, the existing trackers cannot handle out-of-view and re-appearance challenges. Our method provides high quality of saliency map in detection, and accurate scale and position of the target in tracking.

The end!