Deep Global-Relative Networks for End-to-End 6-DoF Visual Localization and Odometry

by   Yimin Lin, et al.

For the autonomous navigation of mobile robots, robust and fast visual localization is a challenging task. Although some end-to-end deep neural networks for 6-DoF Visual Odometry (VO) have been reported with promising results, they are still unable to solve the drift problem in long-range navigation. In this paper, we propose the deep global-relative networks (DGRNets), which is a novel global and relative fusion framework based on Recurrent Convolutional Neural Networks (RCNNs). It is designed to jointly estimate global pose and relative localization from consecutive monocular images. DGRNets include feature extraction sub-networks for discriminative feature selection, RCNNs-type relative pose estimation subnetworks for smoothing the VO trajectory and RCNNs-type global pose regression sub-networks for avoiding the accumulation of pose errors. We also propose two loss functions: the first one consists of Cross Transformation Constraints (CTC) that utilize geometric consistency of the adjacent frames to train a more accurate relative sub-networks, and the second one is composed of CTC and Mean Square Error (MSE) between the predicted pose and ground truth used to train the end-to-end DGRNets. The competitive experiments on indoor Microsoft 7-Scenes and outdoor KITTI dataset show that our DGRNets outperform other learning-based monocular VO methods in terms of pose accuracy.


page 1

page 3

page 4


Deep Auxiliary Learning for Visual Localization and Odometry

Localization is an indispensable component of a robot's autonomy stack t...

ViPR: Visual-Odometry-aided Pose Regression for 6DoF Camera Localization

Visual Odometry (VO) accumulates a positional drift in long-term robot n...

PoseConvGRU: A Monocular Approach for Visual Ego-motion Estimation by Learning

While many visual ego-motion algorithm variants have been proposed in th...

Ultra-low Power Deep Learning-based Monocular Relative Localization Onboard Nano-quadrotors

Precise relative localization is a crucial functional block for swarm ro...

VLocNet++: Deep Multitask Learning for Semantic Visual Localization and Odometry

Visual localization is one of the fundamental enablers of robot autonomy...

DPC-Net: Deep Pose Correction for Visual Localization

We present a novel method to fuse the power of deep networks with the co...

Deep Keypoint-Based Camera Pose Estimation with Geometric Constraints

Estimating relative camera poses from consecutive frames is a fundamenta...

Please sign up or login with your details

Forgot password? Click here to reset