Presentation is loading. Please wait.

Presentation is loading. Please wait.

Flow Separation for Fast and Robust Stereo Odometry [ICRA 2009]

Similar presentations


Presentation on theme: "Flow Separation for Fast and Robust Stereo Odometry [ICRA 2009]"— Presentation transcript:

1 Flow Separation for Fast and Robust Stereo Odometry [ICRA 2009]
Ph.D. Student, Chang-Ryeol Lee June 26, 2013

2 Contents Introduction Preliminary Proposed method Experimental results
What is Visual Odometry (VO)? Why VO? Terminology Brief history of VO Preliminary One-point RANSAN Proposed method Experimental results

3 Introduction: what is Visual Odometry (VO)?
VO is the process of incrementally estimating the pose of the vehicle by examining the changes that motion induces on the images of its onboard cameras input output Image sequence (or video stream) from one or more cameras attached to a moving vehicle Camera trajectory (3D structure is a plus):

4 Introduction: why VO? Contrary to wheel odometry, VO is not affected by wheel slip in uneven terrain or other adverse conditions. More accurate trajectory estimates compared to wheel odometry (relative position error 0.1% − 2%) VO can be used as a complement to wheel odometry GPS inertial measurement units (IMUs) laser odometry In GPS-denied environments, such as underwater and aerial, VO has utmost importance

5 Introduction: terminology
SFM vs. VO VO is a particular case of SFM VO focuses on estimating the 3D motion of the camera sequentially (as a new frame arrives) and in real time. Bundle adjustment can be used (but it’s optional) to refine the local estimate of the trajectory Sometimes SFM is used as a synonym of VO

6 Introduction: history of VO
1996: The term VO was coined by Srinivasan to define motion orientation in honey bees. 1980: First known stereo VO real-time implementation on a robot by Moraveck PhD thesis (NASA/JPL) for Mars rovers using a sliding camera. Moravec invented a predecessor of Harris detector, known as Moravec detector 1980 to 2000: The VO research was dominated by NASA/JPL in preparation of 2004 Mars mission (see papers from Matthies, Olson, etc. From JPL) 2004: VO used on a robot on another planet: Mars rovers Spirit and Opportunity

7 Introduction: history of VO
2004: VO was revived in the academic environment by Nister «Visual Odometry» paper. The term VO became popular. 2004: Based on loopy belief propagation 2004: Using omnidirectional camera : Focus on large-scale issue in outdoor environments 2007: Landmark handling for improving accuracy

8 Introduction: problem
RANSAC for robust model estimation (usually) Nearly degenerate case in RANSAC Correct matches for fundamental matrix computation are small. Matches on a dominant plane are result in homography.

9 Introduction: problem
Reason to occur nearly degenerate case Bad lighting condition Ground surfaces with low texture Motion blur Result: different inliers

10 Preliminary: three-point VO
Procedure 3D points generation by triangulation in first stereo image. 2. Track features of a next frame. Pose estimation of next frame by P3P algorithm with RANSAC. First frame First frame Second frame

11 Preliminary: three-point VO
Triangulate all new feature matches. 4. Repeat from Step 2 First frame Second frame

12 Proposed method Key idea Contributions
Small changes in the camera translation do not influence points which are far away. ⇒ Separate feature points, two-step model estimation Contributions More robust than 3-point VO (nearly degenerate case handling) Faster than 3-point VO (efficiency)

13 Proposed method Procedure Perform sparse stereo and putative matching.
Separate features based on disparity. Recover rotation with two-point RANSAC. Recover translation with one-point RANSAC.

14 Proposed method Sparse stereo and putative matching
Calibrated and rectified images Sparse stereo 1. Feature extraction. 2. Matching in scan line. Putative matching 1. Prediction of vehicle motion by * odometry * previous motion * stationary assumption. 2. Template matching

15 Proposed method Separate features based on disparity 𝜃
Threshold is based on vehicle speed. , where b is baseline, f is focal length , where { 𝑡 𝑥 , 𝑡 𝑦 , 𝑡 𝑧 } are prediction of vehicle motion , where {△𝑢,△𝑣} are maximum allowed pixel error. (0.1~0.5) Translation -> Threshold > Close feature points Translation -> Threshold > Far feature points 𝜃

16 Proposed method Separate features based on disparity
In the case that either far or close feature points only exist. Only far feature points Translation is zero or small Use a minimum number of the closest putative matches Only close feature points There is no such case since we assume that camera translation is small.

17 Proposed method Rotation: two-point RANSAC
Far feature points are not influenced by camera translation. We regard points for rotation estimation as points at infinity. Points at infinity have 0 disparity (same points in left, right image) -> Rotation estimation is based on the direction of points at infinity -> Monocular approach (use only right or left images)

18 Proposed method Rotation: two-point RANSAC
Each measurement contribute 2 constraints Cost function: reprojection error Unknown: rotation 3DOF * 𝑛 is the number of points Require 2 points at least

19 Proposed method Translation: One-point RANSAC
Intuitively, the difference of each 3D points from a single match in two frames is camera translation -> stereo approach (use stereo images)

20 Proposed method Translation: One-point RANSAC
Each measurement contribute 3 constraints By minimizing re-projection error Unknown: translation 3DOF * 𝑛 is the number of points Require 1 point at least

21 Experimental results Robustness

22 Experimental results Speed and accuracy

23 Experimental results Speed and accuracy

24 Reference [1] D. Scaramuzza, F. Fraundorfer. “Visual Odometry [Tutorial]” Robotics & Automation Magazine, IEEE, Vol. 18, No. 4. December [2] Kaess, M., Ni, K., & Dellaert, F. “Flow Separation for Fast and Robust Stereo Odometry”. Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), [3] D. Nister, O. Naroditsky, J. Bergen. “Visual odometry”. Computer Vision and Pattern Recognition, 2004.

25 Thank you!


Download ppt "Flow Separation for Fast and Robust Stereo Odometry [ICRA 2009]"

Similar presentations


Ads by Google