• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

LVID-SLAM:一种基于语义信息的用于动态场景的轻量级视觉惯性同步定位与地图构建技术

LVID-SLAM: A Lightweight Visual-Inertial SLAM for Dynamic Scenes Based on Semantic Information.

作者信息

Wang Shuwen, Hu Qiming, Zhang Xu, Li Wei, Wang Ying, Zheng Enhui

机构信息

School of Mechanical and Electrical Engineering, China Jiliang University, Hangzhou 310018, China.

出版信息

Sensors (Basel). 2025 Jul 1;25(13):4117. doi: 10.3390/s25134117.

DOI:10.3390/s25134117
PMID:40648372
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12251585/
Abstract

Simultaneous Localization and Mapping (SLAM) remains challenging in dynamic environments. Recent approaches combining deep learning with algorithms for dynamic scenes comprise two types: faster, less accurate object detection-based methods and highly accurate, computationally costly instance segmentation-based methods. In addition, maps lacking semantic information hinder robots from understanding their environment and performing complex tasks. This paper presents a lightweight visual-inertial SLAM system. The system is based on the classic ORB-SLAM3 framework, which starts a new thread for object detection and tightly couples the semantic information of object detection with geometric information to remove feature points from dynamic objects. In addition, Inertial Measurement Unit (IMU) data are employed to assist in feature point extraction, thereby compensating for visual pose tracking loss. Finally, a dense octree-based semantic map is constructed by fusing semantic information and visualized using ROS. LVID-SLAM demonstrates excellent pose accuracy and robustness in highly dynamic scenes on the public TUM dataset, with an average ATE reduction of more than 80% compared to ORB-SLAM3. The experimental results demonstrate that LVID-SLAM outperforms other methods in dynamic conditions, offering both real-time capability and robustness.

摘要

同时定位与地图构建(SLAM)在动态环境中仍然具有挑战性。最近将深度学习与动态场景算法相结合的方法包括两类:基于目标检测的方法速度更快但精度较低,以及基于实例分割的方法精度高但计算成本高昂。此外,缺乏语义信息的地图阻碍机器人理解其环境并执行复杂任务。本文提出了一种轻量级视觉惯性SLAM系统。该系统基于经典的ORB-SLAM3框架,它为目标检测启动一个新线程,并将目标检测的语义信息与几何信息紧密结合,以去除动态物体上的特征点。此外,利用惯性测量单元(IMU)数据辅助特征点提取,从而弥补视觉姿态跟踪的损失。最后,通过融合语义信息构建基于八叉树的密集语义地图,并使用ROS进行可视化。在公开的TUM数据集上,LVID-SLAM在高动态场景中展现出出色的姿态精度和鲁棒性,与ORB-SLAM3相比,平均绝对轨迹误差(ATE)降低了80%以上。实验结果表明,LVID-SLAM在动态条件下优于其他方法,兼具实时性和鲁棒性。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d23d/12251585/c5cda48ca964/sensors-25-04117-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d23d/12251585/bba904bb0e08/sensors-25-04117-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d23d/12251585/227cac74072c/sensors-25-04117-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d23d/12251585/551980fe5921/sensors-25-04117-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d23d/12251585/3b994f1e7d71/sensors-25-04117-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d23d/12251585/e62acfad72dd/sensors-25-04117-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d23d/12251585/8ab5513ad604/sensors-25-04117-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d23d/12251585/a1c83c9c7590/sensors-25-04117-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d23d/12251585/f2526358cf16/sensors-25-04117-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d23d/12251585/b5f574130246/sensors-25-04117-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d23d/12251585/cef6241a47ba/sensors-25-04117-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d23d/12251585/c5cda48ca964/sensors-25-04117-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d23d/12251585/bba904bb0e08/sensors-25-04117-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d23d/12251585/227cac74072c/sensors-25-04117-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d23d/12251585/551980fe5921/sensors-25-04117-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d23d/12251585/3b994f1e7d71/sensors-25-04117-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d23d/12251585/e62acfad72dd/sensors-25-04117-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d23d/12251585/8ab5513ad604/sensors-25-04117-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d23d/12251585/a1c83c9c7590/sensors-25-04117-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d23d/12251585/f2526358cf16/sensors-25-04117-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d23d/12251585/b5f574130246/sensors-25-04117-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d23d/12251585/cef6241a47ba/sensors-25-04117-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d23d/12251585/c5cda48ca964/sensors-25-04117-g011.jpg

相似文献

1
LVID-SLAM: A Lightweight Visual-Inertial SLAM for Dynamic Scenes Based on Semantic Information.LVID-SLAM:一种基于语义信息的用于动态场景的轻量级视觉惯性同步定位与地图构建技术
Sensors (Basel). 2025 Jul 1;25(13):4117. doi: 10.3390/s25134117.
2
SGDO-SLAM: A Semantic RGB-D SLAM System with Coarse-to-Fine Dynamic Rejection and Static Weighted Optimization.SGDO-SLAM:一种具有从粗到精动态剔除和静态加权优化的语义RGB-D同步定位与地图构建系统。
Sensors (Basel). 2025 Jun 14;25(12):3734. doi: 10.3390/s25123734.
3
PLY-SLAM: Semantic Visual SLAM Integrating Point-Line Features with YOLOv8-seg in Dynamic Scenes.PLY-SLAM:动态场景中集成点线特征与YOLOv8分割的语义视觉同步定位与地图构建
Sensors (Basel). 2025 Jun 7;25(12):3597. doi: 10.3390/s25123597.
4
Short-Term Memory Impairment短期记忆障碍
5
SGF-SLAM: Semantic Gaussian Filtering SLAM for Urban Road Environments.SGF-SLAM:用于城市道路环境的语义高斯滤波同步定位与地图构建
Sensors (Basel). 2025 Jun 7;25(12):3602. doi: 10.3390/s25123602.
6
LET-SE2-VINS: A Hybrid Optical Flow Framework for Robust Visual-Inertial SLAM.LET-SE2-VINS:一种用于鲁棒视觉惯性同步定位与地图构建的混合光流框架。
Sensors (Basel). 2025 Jun 20;25(13):3837. doi: 10.3390/s25133837.
7
BY-SLAM: Dynamic Visual SLAM System Based on BEBLID and Semantic Information Extraction.BY-SLAM:基于BEBLID和语义信息提取的动态视觉同步定位与地图构建系统
Sensors (Basel). 2024 Jul 19;24(14):4693. doi: 10.3390/s24144693.
8
SEG-SLAM: Dynamic Indoor RGB-D Visual SLAM Integrating Geometric and YOLOv5-Based Semantic Information.SEG-SLAM:集成几何信息与基于YOLOv5的语义信息的动态室内RGB-D视觉同步定位与地图构建
Sensors (Basel). 2024 Mar 25;24(7):2102. doi: 10.3390/s24072102.
9
ADM-SLAM: Accurate and Fast Dynamic Visual SLAM with Adaptive Feature Point Extraction, Deeplabv3pro, and Multi-View Geometry.ADM-SLAM:基于自适应特征点提取、深度卷积神经网络语义分割模型Deeplabv3pro和多视图几何的精确快速动态视觉同步定位与地图构建
Sensors (Basel). 2024 Jun 2;24(11):3578. doi: 10.3390/s24113578.
10
SOLO-SLAM: A Parallel Semantic SLAM Algorithm for Dynamic Scenes.单目 SLAM:一种用于动态场景的并行语义 SLAM 算法。
Sensors (Basel). 2022 Sep 15;22(18):6977. doi: 10.3390/s22186977.

本文引用的文献

1
Direct Sparse Odometry.直接稀疏里程计。
IEEE Trans Pattern Anal Mach Intell. 2018 Mar;40(3):611-625. doi: 10.1109/TPAMI.2017.2658577. Epub 2017 Apr 12.
2
MonoSLAM: real-time single camera SLAM.单目即时定位与地图构建(MonoSLAM):实时单目相机即时定位与地图构建
IEEE Trans Pattern Anal Mach Intell. 2007 Jun;29(6):1052-67. doi: 10.1109/TPAMI.2007.1049.