Neural Networks2022,Vol.15018.DOI:10.1016/j.neunet.2022.03.005

SelfVIO: Self-supervised deep monocular Visual-Inertial Odometry and depth estimation

Almalioglu, Yasin Turan, Mehmet Saputra, Muhamad Risqi U. de Gusmao, Pedro P. B. Markham, Andrew Trigoni, Niki
Neural Networks2022,Vol.15018.DOI:10.1016/j.neunet.2022.03.005

SelfVIO: Self-supervised deep monocular Visual-Inertial Odometry and depth estimation

Almalioglu, Yasin 1Turan, Mehmet 2Saputra, Muhamad Risqi U. 1de Gusmao, Pedro P. B. 1Markham, Andrew 1Trigoni, Niki1
扫码查看

作者信息

  • 1. Comp Sci Dept,Univ Oxford
  • 2. Inst Biomed Engn,Bogazici Univ
  • 折叠

Abstract

In the last decade, numerous supervised deep learning approaches have been proposed for visual- inertial odometry (VIO) and depth map estimation, which require large amounts of labelled data. To overcome the data limitation, self-supervised learning has emerged as a promising alternative that exploits constraints such as geometric and photometric consistency in the scene. In this study, we present a novel self-supervised deep learning-based VIO and depth map recovery approach (SelfVIO) using adversarial training and self-adaptive visual-inertial sensor fusion. SelfVIO learns the joint estimation of 6 degrees-of-freedom (6-DoF) ego-motion and a depth map of the scene from unlabelled monocular RGB image sequences and inertial measurement unit (IMU) readings. The proposed approach is able to perform VIO without requiring IMU intrinsic parameters and/or extrinsic calibration between IMU and the camera. We provide comprehensive quantitative and qualitative evaluations of the proposed framework and compare its performance with state-of-the-art VIO, VO, and visual simultaneous localization and mapping (VSLAM) approaches on the KITTI, EuRoC and Cityscapes datasets. Detailed comparisons prove that SelfVIO outperforms state-of-the-art VIO approaches in terms of pose estimation and depth recovery, making it a promising approach among existing methods in the literature.(c) 2022 The Author(s). Published by Elsevier Ltd. This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/).

Key words

Self-supervised learning/Geometry reconstruction/Machine perception/Generative adversarial networks/Deep sensor fusion/visual-inertial odometry/OPTICAL-FLOW/EGO-MOTION/NETWORK/VISION/SLAM/CALIBRATION/PREDICTION/VERSATILE/FUSION/ROBUST

引用本文复制引用

出版年

2022
Neural Networks

Neural Networks

EISCI
ISSN:0893-6080
被引量23
参考文献量107
段落导航相关论文