• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于改进深度Q网络算法的AGV路径规划研究

Research on AGV Path Planning Based on Improved DQN Algorithm.

作者信息

Xiao Qian, Pan Tengteng, Wang Kexin, Cui Shuoming

机构信息

School of Intelligent Science Information Engineering, Shenyang University, Shenyang 110044, China.

出版信息

Sensors (Basel). 2025 Jul 29;25(15):4685. doi: 10.3390/s25154685.

DOI:10.3390/s25154685
PMID:40807849
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12349690/
Abstract

Traditional deep reinforcement learning methods suffer from slow convergence speeds and poor adaptability in complex environments and are prone to falling into local optima in AGV system applications. To address these issues, in this paper, an adaptive path planning algorithm with an improved Deep Q Network algorithm called the B-PER DQN algorithm is proposed. Firstly, a dynamic temperature adjustment mechanism is constructed, and the temperature parameters in the Boltzmann strategy are adaptively adjusted by analyzing the change trend of the recent reward window. Next, the Priority experience replay mechanism is introduced to improve the training efficiency and task diversity through experience grading sampling and random obstacle configuration. Then, a refined multi-objective reward function is designed, combined with direction guidance, step punishment, and end point reward, to effectively guide the agent in learning an efficient path. Our experimental results show that, compared with other algorithms, the improved algorithm proposed in this paper achieves a higher success rate and faster convergence in the same environment and represents an efficient and adaptive solution for reinforcement learning for path planning in complex environments.

摘要

传统的深度强化学习方法在复杂环境中存在收敛速度慢和适应性差的问题,并且在AGV系统应用中容易陷入局部最优。为了解决这些问题,本文提出了一种改进的深度Q网络算法——B-PER DQN算法的自适应路径规划算法。首先,构建了一种动态温度调整机制,通过分析近期奖励窗口的变化趋势,对玻尔兹曼策略中的温度参数进行自适应调整。其次,引入了优先经验回放机制,通过经验分级采样和随机障碍物配置来提高训练效率和任务多样性。然后,设计了一种精细化的多目标奖励函数,结合方向引导、步长惩罚和终点奖励,有效地引导智能体学习高效路径。实验结果表明,与其他算法相比,本文提出的改进算法在相同环境下具有更高的成功率和更快的收敛速度,是复杂环境中路径规划强化学习的一种高效且自适应的解决方案。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9bf/12349690/1f0aae19b05c/sensors-25-04685-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9bf/12349690/b97e01cf3496/sensors-25-04685-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9bf/12349690/80dd572b2b9a/sensors-25-04685-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9bf/12349690/a1b7223ad619/sensors-25-04685-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9bf/12349690/444f3af7599a/sensors-25-04685-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9bf/12349690/607b10d630fa/sensors-25-04685-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9bf/12349690/e99e0a328f1a/sensors-25-04685-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9bf/12349690/0e48c977ec4d/sensors-25-04685-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9bf/12349690/7a16d3672999/sensors-25-04685-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9bf/12349690/28666443f324/sensors-25-04685-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9bf/12349690/1f0aae19b05c/sensors-25-04685-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9bf/12349690/b97e01cf3496/sensors-25-04685-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9bf/12349690/80dd572b2b9a/sensors-25-04685-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9bf/12349690/a1b7223ad619/sensors-25-04685-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9bf/12349690/444f3af7599a/sensors-25-04685-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9bf/12349690/607b10d630fa/sensors-25-04685-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9bf/12349690/e99e0a328f1a/sensors-25-04685-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9bf/12349690/0e48c977ec4d/sensors-25-04685-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9bf/12349690/7a16d3672999/sensors-25-04685-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9bf/12349690/28666443f324/sensors-25-04685-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9bf/12349690/1f0aae19b05c/sensors-25-04685-g010.jpg

相似文献

1
Research on AGV Path Planning Based on Improved DQN Algorithm.基于改进深度Q网络算法的AGV路径规划研究
Sensors (Basel). 2025 Jul 29;25(15):4685. doi: 10.3390/s25154685.
2
Research of UAV 3D path planning based on improved Dwarf mongoose algorithm with multiple strategies.基于改进的多策略侏儒 mongoose 算法的无人机三维路径规划研究
Sci Rep. 2025 Jul 24;15(1):26979. doi: 10.1038/s41598-025-11492-y.
3
Actor critic with experience replay-based automatic treatment planning for prostate cancer intensity modulated radiotherapy.基于经验回放的演员-评论家算法用于前列腺癌调强放射治疗的自动治疗计划
Med Phys. 2025 Jul;52(7):e17915. doi: 10.1002/mp.17915. Epub 2025 May 31.
4
Prescription of Controlled Substances: Benefits and Risks管制药品的处方:益处与风险
5
Short-Term Memory Impairment短期记忆障碍
6
A multi strategy bidirectional RRT* algorithm for efficient mobile robot path planning.一种用于高效移动机器人路径规划的多策略双向RRT*算法。
Sci Rep. 2025 Aug 12;15(1):29501. doi: 10.1038/s41598-025-13915-2.
7
Automated guided vehicle (AGV) path optimization method based on improved rapidly-exploring random trees.基于改进的快速扩展随机树的自动导引车(AGV)路径优化方法
PeerJ Comput Sci. 2025 Jun 18;11:e2915. doi: 10.7717/peerj-cs.2915. eCollection 2025.
8
Enhanced Q learning and deep reinforcement learning for unmanned combat intelligence planning in adversarial environments.用于对抗环境中无人作战情报规划的增强型Q学习与深度强化学习
Sci Rep. 2025 Aug 4;15(1):28364. doi: 10.1038/s41598-025-13752-3.
9
Application of Multi-Strategy Controlled Rime Algorithm in Path Planning for Delivery Robots.多策略控制rime算法在配送机器人路径规划中的应用
Biomimetics (Basel). 2025 Jul 19;10(7):476. doi: 10.3390/biomimetics10070476.
10
The Lived Experience of Autistic Adults in Employment: A Systematic Search and Synthesis.成年自闭症患者的就业生活经历:系统检索与综述
Autism Adulthood. 2024 Dec 2;6(4):495-509. doi: 10.1089/aut.2022.0114. eCollection 2024 Dec.

本文引用的文献

1
A Study of the Improved A* Algorithm Incorporating Road Factors for Path Planning in Off-Road Emergency Rescue Scenarios.一种结合道路因素的改进A*算法在越野应急救援场景路径规划中的研究
Sensors (Basel). 2024 Aug 30;24(17):5643. doi: 10.3390/s24175643.
2
Enhancing Stability and Performance in Mobile Robot Path Planning with PMR-Dueling DQN Algorithm.基于PMR-决斗深度Q网络算法提升移动机器人路径规划的稳定性与性能
Sensors (Basel). 2024 Feb 27;24(5):1523. doi: 10.3390/s24051523.