Skip to content

ScaleEstimation/MonoScaleEstimation

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

31 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Simple but Effective Scale Estimation for Monocular Visual Odometry in Road Driving Scenarios

Additional experimental results for the paper titled above.

Abstract

In large scale environments, scale drift is a crucial problem of monocular visual SLAM. A common solution is to utilize the camera height, which can be obtained using the reconstructed 3D ground points~(3DGPs) from two successive frames, as prior knowledge.Increasing the number of 3DGPs by using more proceeding frames can be a natural extension of this solution to estimate a more precise camera height. However, merely employing multiple frames based on conventional methods is hard to be directly applicable in a real-world scenario because the vehicle motion and inaccurate feature matching inevitably cause large uncertainty and noisy 3DGPs. In this study, we propose an elaborate method to collect confident 3DGPs from multiple frames for robust scale estimation. First, we gather 3DGP candidates that can be seen in more than a predefined number of frames. To verify the 3DGP candidates, we filter out the 3D points at the exterior of the road region obtained by the deep-learning-based road segmentation model. In addition, we formulate an optimization problem constrained by a simple but effective geometric assumption that the normal vector of the ground plane lies in the null space of a movement vector of the camera center, and provide a closed-form solution. ORB-SLAM with the proposed scale estimation method achieves the average translation error with 1.19% on the KITTI dataset, which outperforms the state-of-the-art conventional monocular visual


Example of ATE results obtained using SEMK and GCSEMK for the sequences 00, 02-10.


Recorded video for sequence 00, 02-10. The trajectory of ground truth, ORB-SLAM without loop closure and GCSEMK are shown in red, black, and blue respectively. The black sparse points are the reconstructed 3D points. The detected feature points on the ground region are shown as red square in the image.

Recorded video for sequence 00

{% include 00.html%}

Recorded video for sequence 02

{% include 02.html%}

Recorded video for sequence 03

will be available soon

Recorded video for sequence 04

will be available soon

Recorded video for sequence 05

will be available soon

Recorded video for sequence 06

will be available soon

Recorded video for sequence 07

{% include 07.html%}

Recorded video for sequence 08

will be available soon

Recorded video for sequence 09

will be available soon

Recorded video for sequence 10

will be available soon


The estimated camere trajectories using SEMK for sequence 00 and 02-10 in the KITTI benchmark

The translation and rotation errors over different distance segment

The translation errors via |K| and delta

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 2

  •  
  •  

Languages