• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

DynaFusion-SLAM:用于推草机器人的自主导航算法的多传感器融合与动态优化

DynaFusion-SLAM: Multi-Sensor Fusion and Dynamic Optimization of Autonomous Navigation Algorithms for Pasture-Pushing Robot.

作者信息

Liu Zhiwei, Fang Jiandong, Zhao Yudong

机构信息

College of Information Engineering, Inner Mongolia University of Technology, Hohhot 010080, China.

Inner Mongolia Key Laboratory of Intelligent Perception and System Engineering, Hohhot 010080, China.

出版信息

Sensors (Basel). 2025 May 28;25(11):3395. doi: 10.3390/s25113395.

DOI:10.3390/s25113395
PMID:40968963
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12157880/
Abstract

Aiming to address the problems of fewer related studies on autonomous navigation algorithms based on multi-sensor fusion in complex scenarios in pastures, lower degrees of fusion, and insufficient cruising accuracy of the operation path in complex outdoor environments, a multimodal autonomous navigation system is proposed based on a loosely coupled architecture of Cartographer-RTAB-Map (real-time appearance-based mapping). Through laser-vision inertial guidance multi-sensor data fusion, the system achieves high-precision mapping and robust path planning in complex scenes. First, comparing the mainstream laser SLAM algorithms (Hector/Gmapping/Cartographer) through simulation experiments, Cartographer is found to have a significant memory efficiency advantage in large-scale scenarios and is thus chosen as the front-end odometer. Secondly, a two-way position optimization mechanism is innovatively designed: (1) When building the map, Cartographer processes the laser with IMU and odometer data to generate mileage estimations, which provide positioning compensation for RTAB-Map. (2) RTAB-Map fuses the depth camera point cloud and laser data, corrects the global position through visual closed-loop detection, and then uses 2D localization to construct a bimodal environment representation containing a 2D raster map and a 3D point cloud, achieving a complete description of the simulated ranch environment and material morphology and constructing a framework for the navigation algorithm of the pushing robot based on the two types of fused data. During navigation, the combination of RTAB-Map's global localization and AMCL's local localization is used to generate a smoother and robust positional attitude by fusing IMU and odometer data through the EKF algorithm. Global path planning is performed using Dijkstra's algorithm and combined with the TEB (Timed Elastic Band) algorithm for local path planning. Finally, experimental validation is performed in a laboratory-simulated pasture environment. The results indicate that when the RTAB-Map algorithm fuses with the multi-source odometry, its performance is significantly improved in the laboratory-simulated ranch scenario, the maximum absolute value of the error of the map measurement size is narrowed from 24.908 cm to 4.456 cm, the maximum absolute value of the relative error is reduced from 6.227% to 2.025%, and the absolute value of the error at each location is significantly reduced. At the same time, the introduction of multi-source mileage fusion can effectively avoid the phenomenon of large-scale offset or drift in the process of map construction. On this basis, the robot constructs a fusion map containing a simulated pasture environment and material patterns. In the navigation accuracy test experiments, our proposed method reduces the root mean square error (RMSE) coefficient by 1.7% and Std by 2.7% compared with that of RTAB-MAP. The RMSE is reduced by 26.7% and Std by 22.8% compared to that of the AMCL algorithm. On this basis, the robot successfully traverses the six preset points, and the measured X and Y directions and the overall position errors of the six points meet the requirements of the pasture-pushing task. The robot successfully returns to the starting point after completing the task of multi-point navigation, achieving autonomous navigation of the robot.

摘要

针对以往在牧场复杂场景下基于多传感器融合的自主导航算法相关研究较少、融合程度较低以及复杂户外环境中作业路径巡航精度不足等问题,提出了一种基于Cartographer - RTAB - Map(基于实时外观的映射)松散耦合架构的多模态自主导航系统。通过激光 - 视觉惯性制导多传感器数据融合,该系统在复杂场景中实现了高精度映射和稳健的路径规划。首先,通过仿真实验比较主流激光同步定位与地图构建(SLAM)算法(Hector/Gmapping/Cartographer),发现Cartographer在大规模场景中具有显著的内存效率优势,因此被选作前端里程计。其次,创新性地设计了一种双向位置优化机制:(1)在构建地图时,Cartographer利用惯性测量单元(IMU)和里程计数据处理激光,生成里程估计,为RTAB - Map提供定位补偿。(2)RTAB - Map融合深度相机点云和激光数据,通过视觉闭环检测校正全局位置,然后使用二维定位构建包含二维光栅地图和三维点云的双模态环境表示,实现对模拟牧场环境和物料形态的完整描述,并基于这两种融合数据构建推料机器人导航算法框架。在导航过程中,利用RTAB - Map的全局定位与自适应蒙特卡洛定位(AMCL)的局部定位相结合,通过扩展卡尔曼滤波(EKF)算法融合IMU和里程计数据,生成更平滑、稳健的位置姿态。使用迪杰斯特拉算法进行全局路径规划,并结合时间弹性带(TEB)算法进行局部路径规划。最后,在实验室模拟牧场环境中进行实验验证。结果表明,当RTAB - Map算法与多源里程计融合时,在实验室模拟牧场场景中其性能显著提升,地图测量尺寸误差的最大绝对值从24.908厘米缩小到4.456厘米,相对误差的最大绝对值从6.227%降低到2.025%,且各位置误差的绝对值均显著降低。同时,引入多源里程融合可有效避免地图构建过程中出现大规模偏移或漂移现象。在此基础上,机器人构建了包含模拟牧场环境和物料图案的融合地图。在导航精度测试实验中,与RTAB - MAP相比,我们提出的方法将均方根误差(RMSE)系数降低了1.7%,标准差(Std)降低了2.7%。与AMCL算法相比,RMSE降低了26.7%,Std降低了22.8%。在此基础上,机器人成功遍历六个预设点,六个点的测量X和Y方向以及整体位置误差均满足牧场推料任务要求。机器人在完成多点导航任务后成功返回起点,实现了机器人的自主导航。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/e110c00954cc/sensors-25-03395-g025.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/c26a00cc81c9/sensors-25-03395-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/edc48dd179b7/sensors-25-03395-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/0de340bb3d93/sensors-25-03395-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/efce56e487a6/sensors-25-03395-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/a4c58125232b/sensors-25-03395-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/e0c3b67d29f4/sensors-25-03395-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/34b2e936faad/sensors-25-03395-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/fdf646d64e91/sensors-25-03395-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/17964b3985d3/sensors-25-03395-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/0c522867948e/sensors-25-03395-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/9c02ceb5ea9e/sensors-25-03395-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/620c2c2b4805/sensors-25-03395-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/5d6ab20bf577/sensors-25-03395-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/abfd7a719e3f/sensors-25-03395-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/3bc95dc30fdd/sensors-25-03395-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/b366e2e78cca/sensors-25-03395-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/43b4e852cf53/sensors-25-03395-g017.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/4bc8d4d16f85/sensors-25-03395-g018.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/4ef6d861e914/sensors-25-03395-g019.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/21740bce312e/sensors-25-03395-g020.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/dab9531ac996/sensors-25-03395-g021a.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/a7c136e5e6b8/sensors-25-03395-g022.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/54ecc8f562b2/sensors-25-03395-g023.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/a96843bab985/sensors-25-03395-g024.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/e110c00954cc/sensors-25-03395-g025.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/c26a00cc81c9/sensors-25-03395-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/edc48dd179b7/sensors-25-03395-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/0de340bb3d93/sensors-25-03395-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/efce56e487a6/sensors-25-03395-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/a4c58125232b/sensors-25-03395-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/e0c3b67d29f4/sensors-25-03395-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/34b2e936faad/sensors-25-03395-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/fdf646d64e91/sensors-25-03395-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/17964b3985d3/sensors-25-03395-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/0c522867948e/sensors-25-03395-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/9c02ceb5ea9e/sensors-25-03395-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/620c2c2b4805/sensors-25-03395-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/5d6ab20bf577/sensors-25-03395-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/abfd7a719e3f/sensors-25-03395-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/3bc95dc30fdd/sensors-25-03395-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/b366e2e78cca/sensors-25-03395-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/43b4e852cf53/sensors-25-03395-g017.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/4bc8d4d16f85/sensors-25-03395-g018.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/4ef6d861e914/sensors-25-03395-g019.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/21740bce312e/sensors-25-03395-g020.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/dab9531ac996/sensors-25-03395-g021a.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/a7c136e5e6b8/sensors-25-03395-g022.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/54ecc8f562b2/sensors-25-03395-g023.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/a96843bab985/sensors-25-03395-g024.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/39a3/12157880/e110c00954cc/sensors-25-03395-g025.jpg

相似文献

1
DynaFusion-SLAM: Multi-Sensor Fusion and Dynamic Optimization of Autonomous Navigation Algorithms for Pasture-Pushing Robot.DynaFusion-SLAM:用于推草机器人的自主导航算法的多传感器融合与动态优化
Sensors (Basel). 2025 May 28;25(11):3395. doi: 10.3390/s25113395.
2
A long-term localization and mapping system for autonomous inspection robots in large-scale environments using 3D LiDAR sensors.一种用于大型环境中自主巡检机器人的基于3D激光雷达传感器的长期定位与建图系统。
PLoS One. 2025 Jul 31;20(7):e0328169. doi: 10.1371/journal.pone.0328169. eCollection 2025.
3
Research on Disaster Environment Map Fusion Construction and Reinforcement Learning Navigation Technology Based on Air-Ground Collaborative Multi-Heterogeneous Robot Systems.基于空地协同多异构机器人系统的灾害环境地图融合构建与强化学习导航技术研究
Sensors (Basel). 2025 Aug 12;25(16):4988. doi: 10.3390/s25164988.
4
Prescription of Controlled Substances: Benefits and Risks管制药品的处方:益处与风险
5
A multi-robot collaborative manipulation framework for dynamic and obstacle-dense environments: integration of deep learning for real-time task execution.一种用于动态和障碍物密集环境的多机器人协作操纵框架:集成深度学习以实现实时任务执行。
Front Robot AI. 2025 Jul 30;12:1585544. doi: 10.3389/frobt.2025.1585544. eCollection 2025.
6
2HR-Net VSLAM: Robust visual SLAM based on dual high-reliability feature matching in dynamic environments.2HR-Net视觉同步定位与地图构建:基于动态环境中双高可靠性特征匹配的稳健视觉同步定位与地图构建。
PLoS One. 2025 Jul 18;20(7):e0328052. doi: 10.1371/journal.pone.0328052. eCollection 2025.
7
Multithreaded Asynchronous Deep Reinforcement Learning With Multisensor Fusion for Robot Collision Avoidance.用于机器人碰撞避免的多传感器融合多线程异步深度强化学习
IEEE Trans Neural Netw Learn Syst. 2025 Sep;36(9):16128-16142. doi: 10.1109/TNNLS.2025.3556438.
8
Anterior Approach Total Ankle Arthroplasty with Patient-Specific Cut Guides.使用患者特异性截骨导向器的前路全踝关节置换术。
JBJS Essent Surg Tech. 2025 Aug 15;15(3). doi: 10.2106/JBJS.ST.23.00027. eCollection 2025 Jul-Sep.
9
Comparison of Two System Identification Approaches for a Four-Wheel Differential Robot Based on Velocity Command Execution.基于速度指令执行的四轮差动机器人两种系统辨识方法的比较
Sensors (Basel). 2025 Jun 5;25(11):3553. doi: 10.3390/s25113553.
10
Short-Term Memory Impairment短期记忆障碍

本文引用的文献

1
Event-Based Visual/Inertial Odometry for UAV Indoor Navigation.用于无人机室内导航的基于事件的视觉/惯性里程计
Sensors (Basel). 2024 Dec 25;25(1):61. doi: 10.3390/s25010061.
2
Research on Positioning Accuracy of Mobile Robot in Indoor Environment Based on Improved RTABMAP Algorithm.基于改进RTABMAP算法的移动机器人在室内环境中的定位精度研究
Sensors (Basel). 2023 Nov 28;23(23):9468. doi: 10.3390/s23239468.
3
Research on Simultaneous localization and mapping Algorithm based on Lidar and IMU.基于激光雷达和惯性测量单元的同步定位与建图算法研究
Math Biosci Eng. 2023 Mar 9;20(5):8954-8974. doi: 10.3934/mbe.2023393.
4
An Approach for Autonomous Feeding Robot Path Planning in Poultry Smart Farm.家禽智能农场中自主喂食机器人路径规划方法
Animals (Basel). 2022 Nov 9;12(22):3089. doi: 10.3390/ani12223089.