• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

DiT-SLAM:基于隐式深度表示和紧密耦合图优化的实时密集视觉惯性同步定位与地图构建

DiT-SLAM: Real-Time Dense Visual-Inertial SLAM with Implicit Depth Representation and Tightly-Coupled Graph Optimization.

作者信息

Zhao Mingle, Zhou Dingfu, Song Xibin, Chen Xiuwan, Zhang Liangjun

机构信息

Institute of Remote Sensing and Geographic Information System, Peking University, Beijing 100871, China.

Robotics and Autonomous Driving Laboratory, Baidu Research, Beijing 100085, China.

出版信息

Sensors (Basel). 2022 Apr 28;22(9):3389. doi: 10.3390/s22093389.

DOI:10.3390/s22093389
PMID:35591079
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC9102487/
Abstract

Recently, generating dense maps in real-time has become a hot research topic in the mobile robotics community, since dense maps can provide more informative and continuous features compared with sparse maps. Implicit depth representation (e.g., the depth code) derived from deep neural networks has been employed in the visual-only or visual-inertial simultaneous localization and mapping (SLAM) systems, which achieve promising performances on both camera motion and local dense geometry estimations from monocular images. However, the existing visual-inertial SLAM systems combined with depth codes are either built on a filter-based SLAM framework, which can only update poses and maps in a relatively small local time window, or based on a loosely-coupled framework, while the prior geometric constraints from the depth estimation network have not been employed for boosting the state estimation. To well address these drawbacks, we propose DiT-SLAM, a novel real-time ense visual-inertial SLAM with mplicit depth representation and ightly-coupled graph optimization. Most importantly, the poses, sparse maps, and low-dimensional depth codes are optimized with the tightly-coupled graph by considering the visual, inertial, and depth residuals simultaneously. Meanwhile, we propose a light-weight monocular depth estimation and completion network, which is combined with attention mechanisms and the conditional variational auto-encoder (CVAE) to predict the uncertainty-aware dense depth maps from more low-dimensional codes. Furthermore, a robust point sampling strategy introducing the spatial distribution of 2D feature points is also proposed to provide geometric constraints in the tightly-coupled optimization, especially for textureless or featureless cases in indoor environments. We evaluate our system on open benchmarks. The proposed methods achieve better performances on both the dense depth estimation and the trajectory estimation compared to the baseline and other systems.

摘要

近年来,实时生成密集地图已成为移动机器人领域的一个热门研究课题,因为与稀疏地图相比,密集地图可以提供更多信息且连续的特征。源自深度神经网络的隐式深度表示(例如深度码)已被应用于仅视觉或视觉惯性同步定位与建图(SLAM)系统中,这些系统在相机运动和从单目图像进行局部密集几何估计方面都取得了不错的性能。然而,现有的结合深度码的视觉惯性SLAM系统要么基于基于滤波器的SLAM框架构建,该框架只能在相对较小的局部时间窗口内更新位姿和地图,要么基于松耦合框架,而深度估计网络的先验几何约束尚未用于提升状态估计。为了很好地解决这些缺点,我们提出了DiT-SLAM,一种具有隐式深度表示和紧耦合图优化的新型实时密集视觉惯性SLAM。最重要的是,通过同时考虑视觉、惯性和深度残差,利用紧耦合图对位姿、稀疏地图和低维深度码进行优化。同时,我们提出了一个轻量级的单目深度估计与完成网络,它结合了注意力机制和条件变分自动编码器(CVAE),以从更多低维码预测具有不确定性感知的密集深度图。此外,还提出了一种引入二维特征点空间分布的鲁棒点采样策略,以在紧耦合优化中提供几何约束,特别是对于室内环境中无纹理或无特征的情况。我们在开放基准上评估了我们的系统。与基线和其他系统相比,所提出的方法在密集深度估计和轨迹估计方面都取得了更好的性能。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915c/9102487/ff43273280b2/sensors-22-03389-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915c/9102487/f3a2b8cd4368/sensors-22-03389-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915c/9102487/aa18c9cb6e9e/sensors-22-03389-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915c/9102487/d4e8d8acdcb6/sensors-22-03389-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915c/9102487/81c4c07d2ab4/sensors-22-03389-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915c/9102487/047c3d460e04/sensors-22-03389-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915c/9102487/76f8cd7f8fee/sensors-22-03389-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915c/9102487/7a5363bbe89d/sensors-22-03389-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915c/9102487/9b15d0633afb/sensors-22-03389-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915c/9102487/d5343dac7d0a/sensors-22-03389-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915c/9102487/73aed18204ad/sensors-22-03389-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915c/9102487/4d88e88dacf7/sensors-22-03389-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915c/9102487/ff43273280b2/sensors-22-03389-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915c/9102487/f3a2b8cd4368/sensors-22-03389-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915c/9102487/aa18c9cb6e9e/sensors-22-03389-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915c/9102487/d4e8d8acdcb6/sensors-22-03389-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915c/9102487/81c4c07d2ab4/sensors-22-03389-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915c/9102487/047c3d460e04/sensors-22-03389-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915c/9102487/76f8cd7f8fee/sensors-22-03389-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915c/9102487/7a5363bbe89d/sensors-22-03389-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915c/9102487/9b15d0633afb/sensors-22-03389-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915c/9102487/d5343dac7d0a/sensors-22-03389-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915c/9102487/73aed18204ad/sensors-22-03389-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915c/9102487/4d88e88dacf7/sensors-22-03389-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915c/9102487/ff43273280b2/sensors-22-03389-g012.jpg

相似文献

1
DiT-SLAM: Real-Time Dense Visual-Inertial SLAM with Implicit Depth Representation and Tightly-Coupled Graph Optimization.DiT-SLAM:基于隐式深度表示和紧密耦合图优化的实时密集视觉惯性同步定位与地图构建
Sensors (Basel). 2022 Apr 28;22(9):3389. doi: 10.3390/s22093389.
2
CVIDS: A Collaborative Localization and Dense Mapping Framework for Multi-Agent Based Visual-Inertial SLAM.CVIDS:一种用于基于多智能体的视觉惯性同步定位与地图构建的协作式定位与密集建图框架。
IEEE Trans Image Process. 2022;31:6562-6576. doi: 10.1109/TIP.2022.3213189. Epub 2022 Oct 21.
3
Visual-Inertial RGB-D SLAM with Encoder Integration of ORB Triangulation and Depth Measurement Uncertainties.结合ORB三角测量与深度测量不确定性编码器集成的视觉惯性RGB-D同步定位与地图构建
Sensors (Basel). 2024 Sep 14;24(18):5964. doi: 10.3390/s24185964.
4
DOT-SLAM: A Stereo Visual Simultaneous Localization and Mapping (SLAM) System with Dynamic Object Tracking Based on Graph Optimization.DOT-SLAM:一种基于图优化的具有动态目标跟踪功能的立体视觉同步定位与地图构建(SLAM)系统。
Sensors (Basel). 2024 Jul 18;24(14):4676. doi: 10.3390/s24144676.
5
Accurate Initial State Estimation in a Monocular Visual-Inertial SLAM System.单目视觉惯性同步定位与地图构建系统中的精确初始状态估计
Sensors (Basel). 2018 Feb 8;18(2):506. doi: 10.3390/s18020506.
6
Depth Completion With Multiple Balanced Bases and Confidence for Dense Monocular SLAM.基于多个平衡基和置信度的深度补全用于密集单目同步定位与地图构建
IEEE Trans Vis Comput Graph. 2025 Sep;31(9):5158-5169. doi: 10.1109/TVCG.2024.3431926.
7
Direct Depth SLAM: Sparse Geometric Feature Enhanced Direct Depth SLAM System for Low-Texture Environments.直接深度 SLAM:用于低纹理环境的稀疏几何特征增强的直接深度 SLAM 系统。
Sensors (Basel). 2018 Oct 6;18(10):3339. doi: 10.3390/s18103339.
8
RC-SLAM: Road Constrained Stereo Visual SLAM System Based on Graph Optimization.RC-SLAM:基于图优化的道路约束立体视觉同步定位与地图构建系统
Sensors (Basel). 2024 Jan 15;24(2):0. doi: 10.3390/s24020536.
9
SD-VIS: A Fast and Accurate Semi-Direct Monocular Visual-Inertial Simultaneous Localization and Mapping (SLAM).SD-VIS:一种快速且精确的半直接单目视觉惯性同步定位与地图构建(SLAM)方法。
Sensors (Basel). 2020 Mar 9;20(5):1511. doi: 10.3390/s20051511.
10
SLAM-based dense surface reconstruction in monocular Minimally Invasive Surgery and its application to Augmented Reality.基于 SLAM 的单目微创手术中密集表面重建及其在增强现实中的应用。
Comput Methods Programs Biomed. 2018 May;158:135-146. doi: 10.1016/j.cmpb.2018.02.006. Epub 2018 Feb 8.

本文引用的文献

1
Learning Guided Convolutional Network for Depth Completion.用于深度补全的学习引导卷积网络。
IEEE Trans Image Process. 2021;30:1116-1129. doi: 10.1109/TIP.2020.3040528. Epub 2020 Dec 15.
2
Towards Robust Monocular Depth Estimation: Mixing Datasets for Zero-Shot Cross-Dataset Transfer.迈向稳健的单目深度估计:混合数据集以实现零样本跨数据集迁移。
IEEE Trans Pattern Anal Mach Intell. 2022 Mar;44(3):1623-1637. doi: 10.1109/TPAMI.2020.3019967. Epub 2022 Feb 3.
3
Direct Sparse Odometry.直接稀疏里程计。
IEEE Trans Pattern Anal Mach Intell. 2018 Mar;40(3):611-625. doi: 10.1109/TPAMI.2017.2658577. Epub 2017 Apr 12.
4
MonoSLAM: real-time single camera SLAM.单目即时定位与地图构建(MonoSLAM):实时单目相机即时定位与地图构建
IEEE Trans Pattern Anal Mach Intell. 2007 Jun;29(6):1052-67. doi: 10.1109/TPAMI.2007.1049.