基于无监督深度学习的单目视觉里程计
Monocular Visual Odometry Based on Unsupervised Deep Learning
白宇 1钟锐 2王奥博 2方浩 2刘建涛3
作者信息
- 1. 中国建筑一局(集团)有限公司,北京 100071;中建市政工程有限公司,北京 100071
- 2. 北京理工大学自动化学院,北京 100081
- 3. 河北林创电子科技有限公司,石家庄 050035
- 折叠
摘要
近年来,视觉实时定位与建图技术(visual simultaneous localization and mapping,VSLAM)技术取得了快速发展.然而,大多数传统的VSLAM系统存在鲁棒性较差的问题而基于深度学习的VSLAM存在精度较低等问题.为了提高VSLAM系统的性能,提出了一种基于无监督深度学习的单目VSLAM.该方法首先设计深度估计网络完成对图像的深度估计.然后,设计位姿估计网络进行相机的位姿估计.最后,使用合理的损失函数保证网络在迭代过程中的有效收敛.在KITTI数据集上验证了该系统的性能.实验结果表明,与SfMLearner相比,绝对轨迹误差(absolute trajectory error,ATE)降低大约50%.与传统的VSLAM系统相比,绝对位姿误差(absolute pose error,APE)的平移部分误差也明显下降且鲁棒性得到提升.
Abstract
VSLAM(visual simultaneous localization and mapping)technology makes rapid progress in recent years.However,most traditional VSLAM systems have poor robustness,while VSLAM based on deep learning have low accuracy.In order to improve the performance of VSLAM systems,an unsupervised deep learning based monocular VSLAM was proposed.The depth estimation net-work was designed to complete the depth estimation of the image and the pose estimation network was applied to estimate the pose of the camera.Finally,a reasonable loss function was used to ensure the effective convergence of the network during the iterative process.The performance of the system is verified on KITTI data set.The results show that ATE(absolute trajectory error)is reduced by 50%compared to SfMLearner.Compared with traditional VSLAM system,the translation part of APE(absolute pose error)is also signifi-cantly reduced and the robustness is also improved.
关键词
同步定位与建图/深度估计网络/位姿估计网络/无监督学习Key words
SLAM robotics/depth estimation network/pose estimation network/unsupervised learning引用本文复制引用
基金项目
国家重点研发计划(2022YFA1004703)
国家自然科学基金(62133002)
出版年
2024