flow separation for fast and robust stereo odometry [icra 2009] ph.d. student, chang-ryeol lee june...
TRANSCRIPT
Flow Separation for Fast and Robust
Stereo Odometry [ICRA 2009]
Ph.D. Student, Chang-Ryeol Lee
June 26, 2013
2
Introduction– What is Visual Odometry (VO)?– Why VO?– Terminology– Brief history of VO
Preliminary– One-point RANSAN
Proposed method
Experimental results
Contents
3
VO is the process of incrementally estimating the pose of the ve-hicle by examining the changes that motion induces on the im-ages of its onboard cameras
Introduction: what is Visual Odometry (VO)?
input output
Image sequence (or video stream)from one or more cameras attached to a moving vehicle
Camera trajectory (3D structure is a plus):
4
Contrary to wheel odometry, VO is not affected by wheel slip in uneven terrain or other adverse conditions.
More accurate trajectory estimates compared to wheel odometry (relative position error 0.1% − 2%)
VO can be used as a complement to – wheel odometry – GPS– inertial measurement units (IMUs)– laser odometry
In GPS-denied environments, such as underwater and aerial, VO has utmost importance
Introduction: why VO?
5
SFM vs. VO– VO is a particular case of SFM– VO focuses on estimating the 3D motion of the camera
sequentially (as a new frame arrives) and in real time.– Bundle adjustment can be used (but it’s optional) to refine the lo-
cal estimate of the trajectory– Sometimes SFM is used as a synonym of VO
Introduction: terminology
6
Introduction: history of VO 1996: The term VO was coined by Srinivasan to define motion
orientation in honey bees.
1980: First known stereo VO real-time implementation on a robot by Moraveck PhD thesis (NASA/JPL) for Mars rovers using a sliding camera. Moravec invented a predecessor of Harris detector, known as Moravec detector
1980 to 2000: The VO research was dominated by NASA/JPL in preparation of 2004 Mars mission (see papers from Matthies, Olson, etc. From JPL)
2004: VO used on a robot on another planet: Mars rovers Spirit and Opportunity
7
2004: VO was revived in the academic environment by Nister «Visual Odometry» paper. The term VO became popular.
2004: Based on loopy belief propagation
2004: Using omnidirectional camera
2006-2007: Focus on large-scale issue in outdoor environments
2007: Landmark handling for improving accuracy
Introduction: history of VO
8
RANSAC for robust model estimation (usually)
Nearly degenerate case in RANSAC– Correct matches for fundamental matrix computation are
small.– Matches on a dominant plane are result in homography.
Introduction: problem
9
Reason to occur nearly degenerate case– Bad lighting condition– Ground surfaces with low texture– Motion blur
Result: different inliers
Introduction: problem
10
Procedure
1. 3D points generation by triangulation in first stereo image.
2. Track features of a next frame. Pose estimation of next frame by P3P algorithm with RANSAC.
Preliminary: three-point VO
First frame
First frameSecond frame
11
3. Triangulate all new feature matches.
4. Repeat from Step 2
Preliminary: three-point VO
First frameSecond frame
12
Key idea– Small changes in the camera translation do not influence
points which are far away. ⇒ Separate feature points, two-step model estimation
Contributions– More robust than 3-point VO (nearly degenerate case
handling)– Faster than 3-point VO (efficiency)
Proposed method
13
Procedure1. Perform sparse stereo and putative matching.2. Separate features based on disparity.3. Recover rotation with two-point RANSAC.4. Recover translation with one-point RANSAC.
Proposed method
14
Sparse stereo and putative matching– Calibrated and rectified images
– Sparse stereo 1. Feature extraction. 2. Matching in scan line.
– Putative matching 1. Prediction of vehicle motion by * odometry * previous motion * stationary assumption. 2. Template matching
Proposed method
15
Separate features based on disparity– Threshold is based on vehicle speed.
, where b is baseline, f is focal length , where {,} are prediction of vehicle motion, where are maximum allowed pixel error. (0.1~0.5)
– Translation -> Threshold -> Close feature points– Translation -> Threshold -> Far feature points
Proposed method
𝜃
16
Separate features based on disparity
– In the case that either far or close feature points only exist.
• Only far feature points– Translation is zero or small– Use a minimum number of the closest putative
matches
• Only close feature points – There is no such case since we assume that camera
translation is small.
Proposed method
17
Rotation: two-point RANSAC– Far feature points are not influenced by camera transla-
tion.– We regard points for rotation estimation as points at in-
finity. – Points at infinity have 0 disparity (same points in left, right
image) -> Rotation estimation is based on the direction of points at in-finity -> Monocular approach (use only right or left images)
Proposed method
18
Rotation: two-point RANSAC
– Each measurement contribute 2 constraints
– Cost function: reprojection error
– Unknown: rotation 3DOF * is the number of points
– Require 2 points at least
Proposed method
( , )R R Ri i iz u v
2ˆ = argmin ( , )i ii
z X R R R
RR R
2 3n
19
Translation: One-point RANSAC– Intuitively, the difference of each 3D points from a single
match in two frames is camera translation -> stereo approach (use stereo images)
Proposed method
20
Translation: One-point RANSAC
– Each measurement contribute 3 constraints
– By minimizing re-projection error
– Unknown: translation 3DOF * is the number of points
– Require 1 point at least
Proposed method
( , , ' )R R R ti i i iz u v u
2ˆˆ = argmin ( , , )i ii
z R X t t t
tt t
3 3n
24
[1] D. Scaramuzza, F. Fraundorfer. “Visual Odometry [Tutorial]” Robotics & Automation Magazine, IEEE, Vol. 18, No. 4. De-cember 2011.
[2] Kaess, M., Ni, K., & Dellaert, F. “Flow Separation for Fast and Ro-bust Stereo Odometry”. Proceedings of the IEEE International Confer-ence on Robotics and Automation (ICRA), 2009.
[3] D. Nister, O. Naroditsky, J. Bergen. “Visual odometry”. Computer Vision and Pattern Recognition, 2004.
Reference