• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

BY-SLAM:基于BEBLID和语义信息提取的动态视觉同步定位与地图构建系统

BY-SLAM: Dynamic Visual SLAM System Based on BEBLID and Semantic Information Extraction.

作者信息

Zhu Daixian, Liu Peixuan, Qiu Qiang, Wei Jiaxin, Gong Ruolin

机构信息

College of Communication and Information Engineering, Xi'an University of Science and Technology, Xi'an 710054, China.

出版信息

Sensors (Basel). 2024 Jul 19;24(14):4693. doi: 10.3390/s24144693.

DOI:10.3390/s24144693
PMID:39066090
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11280888/
Abstract

SLAM is a critical technology for enabling autonomous navigation and positioning in unmanned vehicles. Traditional visual simultaneous localization and mapping algorithms are built upon the assumption of a static scene, overlooking the impact of dynamic targets within real-world environments. Interference from dynamic targets can significantly degrade the system's localization accuracy or even lead to tracking failure. To address these issues, we propose a dynamic visual SLAM system named BY-SLAM, which is based on BEBLID and semantic information extraction. Initially, the BEBLID descriptor is introduced to describe Oriented FAST feature points, enhancing both feature point matching accuracy and speed. Subsequently, FasterNet replaces the backbone network of YOLOv8s to expedite semantic information extraction. By using the results of DBSCAN clustering object detection, a more refined semantic mask is obtained. Finally, by leveraging the semantic mask and epipolar constraints, dynamic feature points are discerned and eliminated, allowing for the utilization of only static feature points for pose estimation and the construction of a dense 3D map that excludes dynamic targets. Experimental evaluations are conducted on both the TUM RGB-D dataset and real-world scenarios and demonstrate the effectiveness of the proposed algorithm at filtering out dynamic targets within the scenes. On average, the localization accuracy for the TUM RGB-D dataset improves by 95.53% compared to ORB-SLAM3. Comparative analyses against classical dynamic SLAM systems further corroborate the improvement in localization accuracy, map readability, and robustness achieved by BY-SLAM.

摘要

SLAM是实现无人驾驶车辆自主导航和定位的关键技术。传统的视觉同步定位与地图构建算法基于静态场景的假设,忽略了现实世界环境中动态目标的影响。动态目标的干扰会显著降低系统的定位精度,甚至导致跟踪失败。为了解决这些问题,我们提出了一种名为BY-SLAM的动态视觉SLAM系统,它基于BEBLID和语义信息提取。首先,引入BEBLID描述符来描述定向FAST特征点,提高特征点匹配的准确性和速度。随后,FasterNet取代了YOLOv8s的骨干网络以加快语义信息提取。通过使用DBSCAN聚类目标检测的结果,获得更精细的语义掩码。最后,利用语义掩码和极线约束,识别并消除动态特征点,从而仅使用静态特征点进行位姿估计,并构建排除动态目标的密集3D地图。在TUM RGB-D数据集和现实场景中进行了实验评估,结果表明所提算法在滤除场景中的动态目标方面是有效的。与ORB-SLAM3相比,TUM RGB-D数据集的定位精度平均提高了95.53%。与经典动态SLAM系统的对比分析进一步证实了BY-SLAM在定位精度、地图可读性和鲁棒性方面的提升。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/108b/11280888/b1d379918486/sensors-24-04693-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/108b/11280888/70a038ea552f/sensors-24-04693-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/108b/11280888/f8b5e4650878/sensors-24-04693-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/108b/11280888/a13abc52acfc/sensors-24-04693-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/108b/11280888/60bfc3c2bbff/sensors-24-04693-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/108b/11280888/d34f368096ad/sensors-24-04693-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/108b/11280888/89d43fa28d3f/sensors-24-04693-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/108b/11280888/e099f0971121/sensors-24-04693-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/108b/11280888/db4009d1dcfc/sensors-24-04693-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/108b/11280888/ec244fe376d5/sensors-24-04693-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/108b/11280888/a409ab489640/sensors-24-04693-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/108b/11280888/9a9577cab43d/sensors-24-04693-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/108b/11280888/705aec690913/sensors-24-04693-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/108b/11280888/b1d379918486/sensors-24-04693-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/108b/11280888/70a038ea552f/sensors-24-04693-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/108b/11280888/f8b5e4650878/sensors-24-04693-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/108b/11280888/a13abc52acfc/sensors-24-04693-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/108b/11280888/60bfc3c2bbff/sensors-24-04693-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/108b/11280888/d34f368096ad/sensors-24-04693-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/108b/11280888/89d43fa28d3f/sensors-24-04693-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/108b/11280888/e099f0971121/sensors-24-04693-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/108b/11280888/db4009d1dcfc/sensors-24-04693-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/108b/11280888/ec244fe376d5/sensors-24-04693-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/108b/11280888/a409ab489640/sensors-24-04693-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/108b/11280888/9a9577cab43d/sensors-24-04693-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/108b/11280888/705aec690913/sensors-24-04693-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/108b/11280888/b1d379918486/sensors-24-04693-g013.jpg

相似文献

1
BY-SLAM: Dynamic Visual SLAM System Based on BEBLID and Semantic Information Extraction.BY-SLAM:基于BEBLID和语义信息提取的动态视觉同步定位与地图构建系统
Sensors (Basel). 2024 Jul 19;24(14):4693. doi: 10.3390/s24144693.
2
Robust visual SLAM algorithm based on target detection and clustering in dynamic scenarios.基于动态场景中目标检测与聚类的鲁棒视觉同步定位与地图构建算法。
Front Neurorobot. 2024 Jul 23;18:1431897. doi: 10.3389/fnbot.2024.1431897. eCollection 2024.
3
SOLO-SLAM: A Parallel Semantic SLAM Algorithm for Dynamic Scenes.单目 SLAM:一种用于动态场景的并行语义 SLAM 算法。
Sensors (Basel). 2022 Sep 15;22(18):6977. doi: 10.3390/s22186977.
4
SEG-SLAM: Dynamic Indoor RGB-D Visual SLAM Integrating Geometric and YOLOv5-Based Semantic Information.SEG-SLAM:集成几何信息与基于YOLOv5的语义信息的动态室内RGB-D视觉同步定位与地图构建
Sensors (Basel). 2024 Mar 25;24(7):2102. doi: 10.3390/s24072102.
5
Semantic visual simultaneous localization and mapping (SLAM) using deep learning for dynamic scenes.使用深度学习的语义视觉同步定位与地图构建(SLAM)用于动态场景。
PeerJ Comput Sci. 2023 Oct 10;9:e1628. doi: 10.7717/peerj-cs.1628. eCollection 2023.
6
A Monocular-Visual SLAM System with Semantic and Optical-Flow Fusion for Indoor Dynamic Environments.一种用于室内动态环境的具有语义和光流融合的单目视觉同步定位与地图构建系统。
Micromachines (Basel). 2022 Nov 17;13(11):2006. doi: 10.3390/mi13112006.
7
YPD-SLAM: A Real-Time VSLAM System for Handling Dynamic Indoor Environments.YPD-SLAM:一种用于处理动态室内环境的实时视觉同步定位与地图构建系统。
Sensors (Basel). 2022 Nov 7;22(21):8561. doi: 10.3390/s22218561.
8
AHY-SLAM: Toward Faster and More Accurate Visual SLAM in Dynamic Scenes Using Homogenized Feature Extraction and Object Detection Method.AHY-SLAM:利用匀质化特征提取和目标检测方法实现动态场景下更快更精确的视觉 SLAM。
Sensors (Basel). 2023 Apr 24;23(9):4241. doi: 10.3390/s23094241.
9
RGB-D Visual SLAM Based on Yolov4-Tiny in Indoor Dynamic Environment.基于Yolov4-Tiny的室内动态环境RGB-D视觉同步定位与地图构建
Micromachines (Basel). 2022 Jan 30;13(2):230. doi: 10.3390/mi13020230.
10
SGC-VSLAM: A Semantic and Geometric Constraints VSLAM for Dynamic Indoor Environments.SGC-VSLAM:一种用于动态室内环境的语义和几何约束视觉同步定位与地图构建方法
Sensors (Basel). 2020 Apr 24;20(8):2432. doi: 10.3390/s20082432.

本文引用的文献

1
A Lightweight Visual Simultaneous Localization and Mapping Method with a High Precision in Dynamic Scenes.一种在动态场景中具有高精度的轻量级视觉同步定位与地图构建方法。
Sensors (Basel). 2023 Nov 19;23(22):9274. doi: 10.3390/s23229274.
2
A Method for Reconstructing Background from RGB-D SLAM in Indoor Dynamic Environments.一种用于室内动态环境中 RGB-D SLAM 的背景重建方法。
Sensors (Basel). 2023 Mar 28;23(7):3529. doi: 10.3390/s23073529.
3
RGB-D SLAM in Dynamic Environments Using Point Correlations.基于点相关性的动态环境中的RGB-D同步定位与地图构建
IEEE Trans Pattern Anal Mach Intell. 2022 Jan;44(1):373-389. doi: 10.1109/TPAMI.2020.3010942. Epub 2021 Dec 7.
4
SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation.SegNet:一种用于图像分割的深度卷积编解码器架构。
IEEE Trans Pattern Anal Mach Intell. 2017 Dec;39(12):2481-2495. doi: 10.1109/TPAMI.2016.2644615. Epub 2017 Jan 2.
5
Discriminative learning of local image descriptors.局部图像描述符的判别式学习。
IEEE Trans Pattern Anal Mach Intell. 2011 Jan;33(1):43-57. doi: 10.1109/TPAMI.2010.54.
6
MonoSLAM: real-time single camera SLAM.单目即时定位与地图构建(MonoSLAM):实时单目相机即时定位与地图构建
IEEE Trans Pattern Anal Mach Intell. 2007 Jun;29(6):1052-67. doi: 10.1109/TPAMI.2007.1049.