• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于强化学习的机器人避障运动规划

Robot movement planning for obstacle avoidance using reinforcement learning.

作者信息

Schneider Linda-Sophie, Peng Junyan, Maier Andreas

机构信息

Pattern Recognition Lab, Friedrich-Alexander-University Erlangen-Nuremberg, Erlangen, Germany.

出版信息

Sci Rep. 2025 Sep 12;15(1):32506. doi: 10.1038/s41598-025-17740-5.

DOI:10.1038/s41598-025-17740-5
PMID:40940370
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12432133/
Abstract

In modern industrial and laboratory environments, robotic arms often operate in complex, cluttered spaces. Ensuring reliable obstacle avoidance and efficient motion planning is therefore essential for safe performance. Motivated by the shortcomings of traditional path planning methods and the growing demand for intelligent automation, we propose a novel reinforcement learning framework that combines a modified artificial potential field (APF) method with the Deep Deterministic Policy Gradient algorithm. Our model is formulated in a continuous environment, which more accurately reflects real-world conditions compared to discrete models. This approach directly addresses the common local optimum issues of conventional APF, enabling the robot arm to navigate complex three-dimensional spaces, optimize its end-effector trajectory, and ensure full-body collision avoidance. Our main contributions include the integration of reinforcement learning factors into the APF framework and the design of a tailored reward mechanism with a compensation term to correct for suboptimal motion directions. This design not only mitigates the inherent limitations of APF in environments with closely spaced obstacles, but also improves performance in both simple and complex scenarios. Extensive experiments show that our method achieves safe and efficient obstacle avoidance with fewer steps and lower energy consumption compared to baseline models, including a TD3-based variant. These results clearly demonstrate the significant potential of our approach to advance robot motion planning in practical applications.

摘要

在现代工业和实验室环境中,机器人手臂常常在复杂、杂乱的空间中运行。因此,确保可靠的避障和高效的运动规划对于安全运行至关重要。鉴于传统路径规划方法的缺点以及对智能自动化的需求不断增长,我们提出了一种新颖的强化学习框架,该框架将改进的人工势场(APF)方法与深度确定性策略梯度算法相结合。我们的模型是在连续环境中构建的,与离散模型相比,它能更准确地反映现实世界的情况。这种方法直接解决了传统APF常见的局部最优问题,使机器人手臂能够在复杂的三维空间中导航,优化其末端执行器轨迹,并确保全身避障。我们的主要贡献包括将强化学习因素整合到APF框架中,以及设计一种带有补偿项的定制奖励机制,以纠正次优运动方向。这种设计不仅减轻了APF在障碍物间距较小的环境中的固有局限性,还提高了在简单和复杂场景中的性能。大量实验表明,与包括基于TD3的变体在内的基线模型相比,我们的方法以更少的步数和更低的能耗实现了安全高效的避障。这些结果清楚地证明了我们的方法在实际应用中推进机器人运动规划的巨大潜力。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6bfb/12432133/416920ba3f93/41598_2025_17740_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6bfb/12432133/8a56f5943de0/41598_2025_17740_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6bfb/12432133/5c272ad2937b/41598_2025_17740_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6bfb/12432133/d55dd51db7b5/41598_2025_17740_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6bfb/12432133/df5e25d2a3bf/41598_2025_17740_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6bfb/12432133/d860b4659e2f/41598_2025_17740_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6bfb/12432133/416920ba3f93/41598_2025_17740_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6bfb/12432133/8a56f5943de0/41598_2025_17740_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6bfb/12432133/5c272ad2937b/41598_2025_17740_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6bfb/12432133/d55dd51db7b5/41598_2025_17740_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6bfb/12432133/df5e25d2a3bf/41598_2025_17740_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6bfb/12432133/d860b4659e2f/41598_2025_17740_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6bfb/12432133/416920ba3f93/41598_2025_17740_Fig6_HTML.jpg

相似文献

1
Robot movement planning for obstacle avoidance using reinforcement learning.基于强化学习的机器人避障运动规划
Sci Rep. 2025 Sep 12;15(1):32506. doi: 10.1038/s41598-025-17740-5.
2
Multithreaded Asynchronous Deep Reinforcement Learning With Multisensor Fusion for Robot Collision Avoidance.用于机器人碰撞避免的多传感器融合多线程异步深度强化学习
IEEE Trans Neural Netw Learn Syst. 2025 Sep;36(9):16128-16142. doi: 10.1109/TNNLS.2025.3556438.
3
Post-pandemic planning for maternity care for local, regional, and national maternity systems across the four nations: a mixed-methods study.针对四个地区的地方、区域和国家孕产妇保健系统的疫情后规划:一项混合方法研究。
Health Soc Care Deliv Res. 2025 Sep;13(35):1-25. doi: 10.3310/HHTE6611.
4
A multi-robot collaborative manipulation framework for dynamic and obstacle-dense environments: integration of deep learning for real-time task execution.一种用于动态和障碍物密集环境的多机器人协作操纵框架:集成深度学习以实现实时任务执行。
Front Robot AI. 2025 Jul 30;12:1585544. doi: 10.3389/frobt.2025.1585544. eCollection 2025.
5
Trajectory tracking and obstacle avoidance in dynamic environments using an improved artificial potential field method.基于改进人工势场法的动态环境下轨迹跟踪与避障
PLoS One. 2025 Jul 10;20(7):e0326879. doi: 10.1371/journal.pone.0326879. eCollection 2025.
6
Prescription of Controlled Substances: Benefits and Risks管制药品的处方:益处与风险
7
Research on Robot Obstacle Avoidance and Generalization Methods Based on Fusion Policy Transfer Learning.基于融合策略迁移学习的机器人避障与泛化方法研究
Biomimetics (Basel). 2025 Jul 25;10(8):493. doi: 10.3390/biomimetics10080493.
8
Design a path - planning strategy for mobile robot in multi-structured environment based on distributional reinforcement learning.基于分布式强化学习设计多结构环境下移动机器人的路径规划策略。
MethodsX. 2025 Aug 7;15:103554. doi: 10.1016/j.mex.2025.103554. eCollection 2025 Dec.
9
Decentralized Multi-Robot Navigation Based on Deep Reinforcement Learning and Trajectory Optimization.基于深度强化学习和轨迹优化的分布式多机器人导航
Biomimetics (Basel). 2025 Jun 4;10(6):366. doi: 10.3390/biomimetics10060366.
10
Integrated neural network framework for multi-object detection and recognition using UAV imagery.用于使用无人机图像进行多目标检测与识别的集成神经网络框架。
Front Neurorobot. 2025 Jul 30;19:1643011. doi: 10.3389/fnbot.2025.1643011. eCollection 2025.

本文引用的文献

1
Deep Reinforcement Learning Based Trajectory Planning Under Uncertain Constraints.基于深度强化学习的不确定约束下轨迹规划
Front Neurorobot. 2022 May 2;16:883562. doi: 10.3389/fnbot.2022.883562. eCollection 2022.