flow separation for fast and robust stereo odometry [icra 2009] ph.d. student, chang-ryeol lee june...

25
Flow Separation for Fast and Robust Stereo Odometry [ICRA 2009] Ph.D. Student, Chang-Ryeol Lee June 26, 2013

Upload: arlene-mckenzie

Post on 30-Dec-2015

217 views

Category:

Documents


3 download

TRANSCRIPT

Flow Separation for Fast and Robust

Stereo Odometry [ICRA 2009]

Ph.D. Student, Chang-Ryeol Lee

June 26, 2013

2

Introduction– What is Visual Odometry (VO)?– Why VO?– Terminology– Brief history of VO

Preliminary– One-point RANSAN

Proposed method

Experimental results

Contents

3

VO is the process of incrementally estimating the pose of the ve-hicle by examining the changes that motion induces on the im-ages of its onboard cameras

Introduction: what is Visual Odometry (VO)?

input output

Image sequence (or video stream)from one or more cameras attached to a moving vehicle

Camera trajectory (3D structure is a plus):

 

 

4

Contrary to wheel odometry, VO is not affected by wheel slip in uneven terrain or other adverse conditions.

More accurate trajectory estimates compared to wheel odometry (relative position error 0.1% − 2%)

VO can be used as a complement to – wheel odometry – GPS– inertial measurement units (IMUs)– laser odometry

In GPS-denied environments, such as underwater and aerial, VO has utmost importance

Introduction: why VO?

5

SFM vs. VO– VO is a particular case of SFM– VO focuses on estimating the 3D motion of the camera

sequentially (as a new frame arrives) and in real time.– Bundle adjustment can be used (but it’s optional) to refine the lo-

cal estimate of the trajectory– Sometimes SFM is used as a synonym of VO

Introduction: terminology

6

Introduction: history of VO 1996: The term VO was coined by Srinivasan to define motion

orientation in honey bees.

1980: First known stereo VO real-time implementation on a robot by Moraveck PhD thesis (NASA/JPL) for Mars rovers using a sliding camera. Moravec invented a predecessor of Harris detector, known as Moravec detector

1980 to 2000: The VO research was dominated by NASA/JPL in preparation of 2004 Mars mission (see papers from Matthies, Olson, etc. From JPL)

2004: VO used on a robot on another planet: Mars rovers Spirit and Opportunity

7

2004: VO was revived in the academic environment by Nister «Visual Odometry» paper. The term VO became popular.

2004: Based on loopy belief propagation

2004: Using omnidirectional camera

2006-2007: Focus on large-scale issue in outdoor environments

2007: Landmark handling for improving accuracy

Introduction: history of VO

8

RANSAC for robust model estimation (usually)

Nearly degenerate case in RANSAC– Correct matches for fundamental matrix computation are

small.– Matches on a dominant plane are result in homography.

Introduction: problem

9

Reason to occur nearly degenerate case– Bad lighting condition– Ground surfaces with low texture– Motion blur

Result: different inliers

Introduction: problem

10

Procedure

1. 3D points generation by triangulation in first stereo image.

2. Track features of a next frame. Pose estimation of next frame by P3P algorithm with RANSAC.

Preliminary: three-point VO

First frame

First frameSecond frame

11

3. Triangulate all new feature matches.

4. Repeat from Step 2

Preliminary: three-point VO

First frameSecond frame

12

Key idea– Small changes in the camera translation do not influence

points which are far away. ⇒ Separate feature points, two-step model estimation

Contributions– More robust than 3-point VO (nearly degenerate case

handling)– Faster than 3-point VO (efficiency)

Proposed method

13

Procedure1. Perform sparse stereo and putative matching.2. Separate features based on disparity.3. Recover rotation with two-point RANSAC.4. Recover translation with one-point RANSAC.

Proposed method

14

Sparse stereo and putative matching– Calibrated and rectified images

– Sparse stereo 1. Feature extraction. 2. Matching in scan line.

– Putative matching 1. Prediction of vehicle motion by * odometry * previous motion * stationary assumption. 2. Template matching

Proposed method

15

Separate features based on disparity– Threshold is based on vehicle speed.

, where b is baseline, f is focal length , where {,} are prediction of vehicle motion, where are maximum allowed pixel error. (0.1~0.5)

– Translation -> Threshold -> Close feature points– Translation -> Threshold -> Far feature points

Proposed method

𝜃

16

Separate features based on disparity

– In the case that either far or close feature points only exist.

• Only far feature points– Translation is zero or small– Use a minimum number of the closest putative

matches

• Only close feature points – There is no such case since we assume that camera

translation is small.

Proposed method

17

Rotation: two-point RANSAC– Far feature points are not influenced by camera transla-

tion.– We regard points for rotation estimation as points at in-

finity. – Points at infinity have 0 disparity (same points in left, right

image) -> Rotation estimation is based on the direction of points at in-finity -> Monocular approach (use only right or left images)

Proposed method

18

Rotation: two-point RANSAC

– Each measurement contribute 2 constraints

– Cost function: reprojection error

– Unknown: rotation 3DOF * is the number of points

– Require 2 points at least

Proposed method

( , )R R Ri i iz u v

2ˆ = argmin ( , )i ii

z X R R R

RR R

2 3n

19

Translation: One-point RANSAC– Intuitively, the difference of each 3D points from a single

match in two frames is camera translation -> stereo approach (use stereo images)

Proposed method

20

Translation: One-point RANSAC

– Each measurement contribute 3 constraints

– By minimizing re-projection error

– Unknown: translation 3DOF * is the number of points

– Require 1 point at least

Proposed method

( , , ' )R R R ti i i iz u v u

2ˆˆ = argmin ( , , )i ii

z R X t t t

tt t

3 3n

21

Robustness

Experimental results

22

Speed and accuracy

Experimental results

23

Speed and accuracy

Experimental results

24

[1] D. Scaramuzza, F. Fraundorfer. “Visual Odometry [Tutorial]” Robotics & Automation Magazine, IEEE, Vol. 18, No. 4. De-cember 2011.

[2] Kaess, M., Ni, K., & Dellaert, F. “Flow Separation for Fast and Ro-bust Stereo Odometry”. Proceedings of the IEEE International Confer-ence on Robotics and Automation (ICRA), 2009.

[3] D. Nister, O. Naroditsky, J. Bergen. “Visual odometry”. Computer Vision and Pattern Recognition, 2004.

Reference

25

Thank you!