Huang Xiaotao, Chen Xingbin, Zhang Ning, He Hongjie, Feng Sang
School of Electromechanical Engineering, Guangdong University of Technology, Guangzhou 510006, China.
Guangdong Productivity Promotion Center, Guangzhou 510075, China.
Sensors (Basel). 2024 Jun 2;24(11):3578. doi: 10.3390/s24113578.
Visual Simultaneous Localization and Mapping (V-SLAM) plays a crucial role in the development of intelligent robotics and autonomous navigation systems. However, it still faces significant challenges in handling highly dynamic environments. The prevalent method currently used for dynamic object recognition in the environment is deep learning. However, models such as Yolov5 and Mask R-CNN require significant computational resources, which limits their potential in real-time applications due to hardware and time constraints. To overcome this limitation, this paper proposes ADM-SLAM, a visual SLAM system designed for dynamic environments that builds upon the ORB-SLAM2. This system integrates efficient adaptive feature point homogenization extraction, lightweight deep learning semantic segmentation based on an improved DeepLabv3, and multi-view geometric segmentation. It optimizes keyframe extraction, segments potential dynamic objects using contextual information with the semantic segmentation network, and detects the motion states of dynamic objects using multi-view geometric methods, thereby eliminating dynamic interference points. The results indicate that ADM-SLAM outperforms ORB-SLAM2 in dynamic environments, especially in high-dynamic scenes, where it achieves up to a 97% reduction in Absolute Trajectory Error (ATE). In various highly dynamic test sequences, ADM-SLAM outperforms DS-SLAM and DynaSLAM in terms of real-time performance and accuracy, proving its excellent adaptability.
视觉同步定位与建图(V-SLAM)在智能机器人和自主导航系统的发展中起着至关重要的作用。然而,它在处理高度动态环境时仍面临重大挑战。当前用于环境中动态物体识别的普遍方法是深度学习。然而,诸如Yolov5和Mask R-CNN等模型需要大量计算资源,由于硬件和时间限制,这限制了它们在实时应用中的潜力。为了克服这一限制,本文提出了ADM-SLAM,这是一种基于ORB-SLAM2设计的用于动态环境的视觉SLAM系统。该系统集成了高效的自适应特征点均匀化提取、基于改进的DeepLabv3的轻量级深度学习语义分割以及多视图几何分割。它优化关键帧提取,使用语义分割网络的上下文信息对潜在动态物体进行分割,并使用多视图几何方法检测动态物体的运动状态,从而消除动态干扰点。结果表明,ADM-SLAM在动态环境中优于ORB-SLAM2,特别是在高动态场景中,其绝对轨迹误差(ATE)降低了高达97%。在各种高动态测试序列中,ADM-SLAM在实时性能和准确性方面优于DS-SLAM和DynaSLAM,证明了其出色的适应性。