• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

改进的双深度Q网络算法应用于六足机器人多维环境路径规划

Improved Double Deep Q-Network Algorithm Applied to Multi-Dimensional Environment Path Planning of Hexapod Robots.

作者信息

Chen Liuhongxu, Wang Qibiao, Deng Chao, Xie Bo, Tuo Xianguo, Jiang Gang

机构信息

School of Computer Science and Engineering, Sichuan University of Science and Engineering, Zigong 643000, China.

School of Physics and Electronic Engineering, Sichuan University of Science and Engineering, Zigong 643000, China.

出版信息

Sensors (Basel). 2024 Mar 23;24(7):2061. doi: 10.3390/s24072061.

DOI:10.3390/s24072061
PMID:38610271
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11013983/
Abstract

Detecting transportation pipeline leakage points within chemical plants is difficult due to complex pathways, multi-dimensional survey points, and highly dynamic scenarios. However, hexapod robots' maneuverability and adaptability make it an ideal candidate for conducting surveys across different planes. The path-planning problem of hexapod robots in multi-dimensional environments is a significant challenge, especially when identifying suitable transition points and planning shorter paths to reach survey points while traversing multi-level environments. This study proposes a Particle Swarm Optimization (PSO)-guided Double Deep Q-Network (DDQN) approach, namely, the PSO-guided DDQN (PG-DDQN) algorithm, for solving this problem. The proposed algorithm incorporates the PSO algorithm to supplant the traditional random selection strategy, and the data obtained from this guided approach are subsequently employed to train the DDQN neural network. The multi-dimensional random environment is abstracted into localized maps comprising current and next level planes. Comparative experiments were performed with PG-DDQN, standard DQN, and standard DDQN to evaluate the algorithm's performance by using multiple randomly generated localized maps. After testing each iteration, each algorithm obtained the total reward values and completion times. The results demonstrate that PG-DDQN exhibited faster convergence under an equivalent iteration count. Compared with standard DQN and standard DDQN, reductions in path-planning time of at least 33.94% and 42.60%, respectively, were observed, significantly improving the robot's mobility. Finally, the PG-DDQN algorithm was integrated with sensors onto a hexapod robot, and validation was performed through Gazebo simulations and Experiment. The results show that controlling hexapod robots by applying PG-DDQN provides valuable insights for path planning to reach transportation pipeline leakage points within chemical plants.

摘要

由于化工厂内输送管道的路径复杂、测量点多维度且场景高度动态,检测输送管道泄漏点具有挑战性。然而,六足机器人的机动性和适应性使其成为在不同平面进行检测的理想选择。六足机器人在多维度环境中的路径规划问题是一项重大挑战,尤其是在识别合适的过渡点并规划更短路径以在穿越多层环境时到达测量点方面。本研究提出一种粒子群优化(PSO)引导的双深度Q网络(DDQN)方法,即PSO引导的DDQN(PG-DDQN)算法来解决此问题。所提出的算法纳入PSO算法以取代传统的随机选择策略,并且从这种引导方法获得的数据随后用于训练DDQN神经网络。将多维度随机环境抽象为包含当前和下一层平面的局部地图。使用多个随机生成的局部地图,对PG-DDQN、标准DQN和标准DDQN进行了对比实验,以评估算法的性能。在每次迭代测试后,各算法获得总奖励值和完成时间。结果表明,在等效迭代次数下,PG-DDQN收敛速度更快。与标准DQN和标准DDQN相比,路径规划时间分别至少减少了33.94%和42.60%,显著提高了机器人的机动性。最后,将PG-DDQN算法与传感器集成到六足机器人上,并通过Gazebo模拟和实验进行验证。结果表明,应用PG-DDQN控制六足机器人为化工厂内输送管道泄漏点的路径规划提供了有价值的见解。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/2ae5740f3e60/sensors-24-02061-g023.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/3a874e2b903d/sensors-24-02061-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/302739e2f560/sensors-24-02061-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/23c07c809f7f/sensors-24-02061-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/3e096f69175b/sensors-24-02061-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/e290e52805b5/sensors-24-02061-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/3c1494ad1805/sensors-24-02061-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/e7df89a67c4d/sensors-24-02061-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/84ce317b4324/sensors-24-02061-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/a9d259175e16/sensors-24-02061-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/0febc7aac2e6/sensors-24-02061-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/b09230465730/sensors-24-02061-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/4a0f56aeda82/sensors-24-02061-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/100c06b7c390/sensors-24-02061-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/9d3749937644/sensors-24-02061-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/09d08908e2cd/sensors-24-02061-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/05a4c09b0ab2/sensors-24-02061-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/d874b0dbfd47/sensors-24-02061-g017.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/fe93ff29175f/sensors-24-02061-g018.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/a0a9287f99e5/sensors-24-02061-g019.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/b26de7643e53/sensors-24-02061-g020.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/9ade9d470246/sensors-24-02061-g021.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/6b58b3326522/sensors-24-02061-g022.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/2ae5740f3e60/sensors-24-02061-g023.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/3a874e2b903d/sensors-24-02061-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/302739e2f560/sensors-24-02061-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/23c07c809f7f/sensors-24-02061-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/3e096f69175b/sensors-24-02061-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/e290e52805b5/sensors-24-02061-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/3c1494ad1805/sensors-24-02061-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/e7df89a67c4d/sensors-24-02061-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/84ce317b4324/sensors-24-02061-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/a9d259175e16/sensors-24-02061-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/0febc7aac2e6/sensors-24-02061-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/b09230465730/sensors-24-02061-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/4a0f56aeda82/sensors-24-02061-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/100c06b7c390/sensors-24-02061-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/9d3749937644/sensors-24-02061-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/09d08908e2cd/sensors-24-02061-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/05a4c09b0ab2/sensors-24-02061-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/d874b0dbfd47/sensors-24-02061-g017.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/fe93ff29175f/sensors-24-02061-g018.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/a0a9287f99e5/sensors-24-02061-g019.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/b26de7643e53/sensors-24-02061-g020.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/9ade9d470246/sensors-24-02061-g021.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/6b58b3326522/sensors-24-02061-g022.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c526/11013983/2ae5740f3e60/sensors-24-02061-g023.jpg

相似文献

1
Improved Double Deep Q-Network Algorithm Applied to Multi-Dimensional Environment Path Planning of Hexapod Robots.改进的双深度Q网络算法应用于六足机器人多维环境路径规划
Sensors (Basel). 2024 Mar 23;24(7):2061. doi: 10.3390/s24072061.
2
Enhancing Stability and Performance in Mobile Robot Path Planning with PMR-Dueling DQN Algorithm.基于PMR-决斗深度Q网络算法提升移动机器人路径规划的稳定性与性能
Sensors (Basel). 2024 Feb 27;24(5):1523. doi: 10.3390/s24051523.
3
The Path Planning of Mobile Robot by Neural Networks and Hierarchical Reinforcement Learning.基于神经网络和分层强化学习的移动机器人路径规划
Front Neurorobot. 2020 Oct 2;14:63. doi: 10.3389/fnbot.2020.00063. eCollection 2020.
4
Improved Robot Path Planning Method Based on Deep Reinforcement Learning.基于深度强化学习的改进型机器人路径规划方法。
Sensors (Basel). 2023 Jun 15;23(12):5622. doi: 10.3390/s23125622.
5
A Mapless Local Path Planning Approach Using Deep Reinforcement Learning Framework.基于深度强化学习框架的无地图局部路径规划方法。
Sensors (Basel). 2023 Feb 10;23(4):2036. doi: 10.3390/s23042036.
6
A Tandem Robotic Arm Inverse Kinematic Solution Based on an Improved Particle Swarm Algorithm.一种基于改进粒子群算法的串联机器人手臂逆运动学求解方法。
Front Bioeng Biotechnol. 2022 May 19;10:832829. doi: 10.3389/fbioe.2022.832829. eCollection 2022.
7
Path Planning Research of a UAV Base Station Searching for Disaster Victims' Location Information Based on Deep Reinforcement Learning.基于深度强化学习的无人机基站搜索灾难受害者位置信息路径规划研究
Entropy (Basel). 2022 Dec 2;24(12):1767. doi: 10.3390/e24121767.
8
Distributed Non-Communicating Multi-Robot Collision Avoidance via Map-Based Deep Reinforcement Learning.基于地图的深度强化学习实现分布式非通信多机器人避碰
Sensors (Basel). 2020 Aug 27;20(17):4836. doi: 10.3390/s20174836.
9
Mapless Path Planning for Mobile Robot Based on Improved Deep Deterministic Policy Gradient Algorithm.基于改进深度确定性策略梯度算法的移动机器人无地图路径规划
Sensors (Basel). 2024 Aug 30;24(17):5667. doi: 10.3390/s24175667.
10
SLP-Improved DDPG Path-Planning Algorithm for Mobile Robot in Large-Scale Dynamic Environment.基于 SLP 的改进 DDPG 路径规划算法在大规模动态环境下的移动机器人应用
Sensors (Basel). 2023 Mar 28;23(7):3521. doi: 10.3390/s23073521.

引用本文的文献

1
Path planning of mobile robot based on improved double deep Q-network algorithm.基于改进双深度Q网络算法的移动机器人路径规划
Front Neurorobot. 2025 Feb 13;19:1512953. doi: 10.3389/fnbot.2025.1512953. eCollection 2025.

本文引用的文献

1
Human-level control through deep reinforcement learning.通过深度强化学习实现人类水平的控制。
Nature. 2015 Feb 26;518(7540):529-33. doi: 10.1038/nature14236.