• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

一种基于多尺度建模的无监督单目视觉里程计

An Unsupervised Monocular Visual Odometry Based on Multi-Scale Modeling.

作者信息

Zhi Henghui, Yin Chenyang, Li Huibin, Pang Shanmin

机构信息

School of Mathematics and Statistics, Xi'an Jiaotong University, Xi'an 710049, China.

School of Software Engineering, Xi'an Jiaotong University, Xi'an 710049, China.

出版信息

Sensors (Basel). 2022 Jul 11;22(14):5193. doi: 10.3390/s22145193.

DOI:10.3390/s22145193
PMID:35890873
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC9323830/
Abstract

Unsupervised deep learning methods have shown great success in jointly estimating camera pose and depth from monocular videos. However, previous methods mostly ignore the importance of multi-scale information, which is crucial for pose estimation and depth estimation, especially when the motion pattern is changed. This article proposes an unsupervised framework for monocular visual odometry (VO) that can model multi-scale information. The proposed method utilizes densely linked atrous convolutions to increase the receptive field size without losing image information, and adopts a non-local self-attention mechanism to effectively model the long-range dependency. Both of them can model objects of different scales in the image, thereby improving the accuracy of VO, especially in rotating scenes. Extensive experiments on the KITTI dataset have shown that our approach is competitive with other state-of-the-art unsupervised learning-based monocular methods and is comparable to supervised or model-based methods. In particular, we have achieved state-of-the-art results on rotation estimation.

摘要

无监督深度学习方法在从单目视频联合估计相机位姿和深度方面已取得巨大成功。然而,先前的方法大多忽略了多尺度信息的重要性,而多尺度信息对于位姿估计和深度估计至关重要,尤其是在运动模式发生变化时。本文提出了一种用于单目视觉里程计(VO)的无监督框架,该框架可以对多尺度信息进行建模。所提出的方法利用密集连接的空洞卷积来增加感受野大小而不丢失图像信息,并采用非局部自注意力机制来有效地对长距离依赖性进行建模。它们都可以对图像中不同尺度的物体进行建模,从而提高VO的准确性,尤其是在旋转场景中。在KITTI数据集上进行的大量实验表明,我们的方法与其他基于无监督学习的最先进单目方法具有竞争力,并且与基于监督或基于模型的方法相当。特别是,我们在旋转估计方面取得了最先进的结果。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a347/9323830/31425d9ff973/sensors-22-05193-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a347/9323830/55470115425f/sensors-22-05193-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a347/9323830/45f3d7295322/sensors-22-05193-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a347/9323830/1285730a9cec/sensors-22-05193-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a347/9323830/aacdf54db4b7/sensors-22-05193-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a347/9323830/4567b6e761bf/sensors-22-05193-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a347/9323830/968d4ca0813d/sensors-22-05193-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a347/9323830/31425d9ff973/sensors-22-05193-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a347/9323830/55470115425f/sensors-22-05193-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a347/9323830/45f3d7295322/sensors-22-05193-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a347/9323830/1285730a9cec/sensors-22-05193-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a347/9323830/aacdf54db4b7/sensors-22-05193-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a347/9323830/4567b6e761bf/sensors-22-05193-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a347/9323830/968d4ca0813d/sensors-22-05193-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a347/9323830/31425d9ff973/sensors-22-05193-g007.jpg

相似文献

1
An Unsupervised Monocular Visual Odometry Based on Multi-Scale Modeling.一种基于多尺度建模的无监督单目视觉里程计
Sensors (Basel). 2022 Jul 11;22(14):5193. doi: 10.3390/s22145193.
2
RAUM-VO: Rotational Adjusted Unsupervised Monocular Visual Odometry.RAUM-VO:旋转调整无监督单目视觉里程计。
Sensors (Basel). 2022 Mar 30;22(7):2651. doi: 10.3390/s22072651.
3
Unsupervised monocular visual odometry via combining instance and RGB information.基于实例和 RGB 信息融合的无监督单目视觉里程计。
Appl Opt. 2022 May 1;61(13):3793-3803. doi: 10.1364/AO.452378.
4
Masked GAN for Unsupervised Depth and Pose Prediction With Scale Consistency.用于具有尺度一致性的无监督深度和姿态预测的掩码生成对抗网络
IEEE Trans Neural Netw Learn Syst. 2021 Dec;32(12):5392-5403. doi: 10.1109/TNNLS.2020.3044181. Epub 2021 Nov 30.
5
Stereo Visual Odometry Pose Correction through Unsupervised Deep Learning.通过无监督深度学习进行立体视觉里程计位姿校正。
Sensors (Basel). 2021 Jul 11;21(14):4735. doi: 10.3390/s21144735.
6
Unsupervised Estimation of Monocular Depth and VO in Dynamic Environments via Hybrid Masks.通过混合掩码对动态环境中的单目深度和视觉里程计进行无监督估计。
IEEE Trans Neural Netw Learn Syst. 2022 May;33(5):2023-2033. doi: 10.1109/TNNLS.2021.3100895. Epub 2022 May 2.
7
SelfVIO: Self-supervised deep monocular Visual-Inertial Odometry and depth estimation.SelfVIO:自监督深度单目视觉惯性里程计和深度估计。
Neural Netw. 2022 Jun;150:119-136. doi: 10.1016/j.neunet.2022.03.005. Epub 2022 Mar 10.
8
Unsupervised Monocular Depth and Camera Pose Estimation with Multiple Masks and Geometric Consistency Constraints.基于多掩码和几何一致性约束的无监督单目深度与相机位姿估计
Sensors (Basel). 2023 Jun 4;23(11):5329. doi: 10.3390/s23115329.
9
WPO-Net: Windowed Pose Optimization Network for Monocular Visual Odometry Estimation.WPO-Net:用于单目视觉里程计估计的窗口姿态优化网络。
Sensors (Basel). 2021 Dec 6;21(23):8155. doi: 10.3390/s21238155.
10
Unsupervised Monocular Visual Odometry for Fast-Moving Scenes Based on Optical Flow Network with Feature Point Matching Constraint.基于特征点匹配约束的光流网络的快速运动场景无监督单目视觉里程计。
Sensors (Basel). 2022 Dec 9;22(24):9647. doi: 10.3390/s22249647.

引用本文的文献

1
Advanced Monocular Outdoor Pose Estimation in Autonomous Systems: Leveraging Optical Flow, Depth Estimation, and Semantic Segmentation with Dynamic Object Removal.自主系统中的高级单目户外姿态估计:利用光流、深度估计和语义分割去除动态物体
Sensors (Basel). 2024 Dec 17;24(24):8040. doi: 10.3390/s24248040.

本文引用的文献

1
WPO-Net: Windowed Pose Optimization Network for Monocular Visual Odometry Estimation.WPO-Net:用于单目视觉里程计估计的窗口姿态优化网络。
Sensors (Basel). 2021 Dec 6;21(23):8155. doi: 10.3390/s21238155.
2
Leveraging Deep Learning for Visual Odometry Using Optical Flow.利用光流的深度学习进行视觉里程计。
Sensors (Basel). 2021 Feb 12;21(4):1313. doi: 10.3390/s21041313.
3
DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs.DeepLab:基于深度卷积网络、空洞卷积和全连接条件随机场的语义图像分割。
IEEE Trans Pattern Anal Mach Intell. 2018 Apr;40(4):834-848. doi: 10.1109/TPAMI.2017.2699184. Epub 2017 Apr 27.
4
Direct Sparse Odometry.直接稀疏里程计。
IEEE Trans Pattern Anal Mach Intell. 2018 Mar;40(3):611-625. doi: 10.1109/TPAMI.2017.2658577. Epub 2017 Apr 12.
5
Image quality assessment: from error visibility to structural similarity.图像质量评估:从误差可见性到结构相似性。
IEEE Trans Image Process. 2004 Apr;13(4):600-12. doi: 10.1109/tip.2003.819861.