• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

CLSQL:基于连续局部搜索策略的移动机器人路径规划改进Q学习算法

CLSQL: Improved Q-Learning Algorithm Based on Continuous Local Search Policy for Mobile Robot Path Planning.

作者信息

Ma Tian, Lyu Jiahao, Yang Jiayi, Xi Runtao, Li Yuancheng, An Jinpeng, Li Chao

机构信息

College of Computer Science and Technology, Xi'an University of Science and Technology, Xi'an 710054, China.

出版信息

Sensors (Basel). 2022 Aug 8;22(15):5910. doi: 10.3390/s22155910.

DOI:10.3390/s22155910
PMID:35957467
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC9371426/
Abstract

How to generate the path planning of mobile robots quickly is a problem in the field of robotics. The Q-learning(QL) algorithm has recently become increasingly used in the field of mobile robot path planning. However, its selection policy is blind in most cases in the early search process, which slows down the convergence of optimal solutions, especially in a complex environment. Therefore, in this paper, we propose a continuous local search Q-Learning (CLSQL) algorithm to solve these problems and ensure the quality of the planned path. First, the global environment is gradually divided into independent local environments. Then, the intermediate points are searched in each local environment with prior knowledge. After that, the search between each intermediate point is realized to reach the destination point. At last, by comparing other RL-based algorithms, the proposed method improves the convergence speed and computation time while ensuring the optimal path.

摘要

如何快速生成移动机器人的路径规划是机器人技术领域的一个问题。Q学习(QL)算法最近在移动机器人路径规划领域的应用越来越广泛。然而,在早期搜索过程中,其选择策略在大多数情况下是盲目性的,这减缓了最优解的收敛速度,尤其是在复杂环境中。因此,在本文中,我们提出了一种连续局部搜索Q学习(CLSQL)算法来解决这些问题,并确保规划路径的质量。首先,将全局环境逐步划分为独立的局部环境。然后,利用先验知识在每个局部环境中搜索中间点。之后,实现各中间点之间的搜索以到达目标点。最后,通过与其他基于强化学习的算法进行比较,该方法在确保最优路径的同时提高了收敛速度和计算时间。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/ff29a3f53e35/sensors-22-05910-g017.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/dfba416e1e82/sensors-22-05910-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/1e639d3e42da/sensors-22-05910-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/0ca40cb3cba5/sensors-22-05910-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/455f67c87226/sensors-22-05910-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/bfc6053989b3/sensors-22-05910-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/533524974360/sensors-22-05910-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/8119fe092d5f/sensors-22-05910-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/47d6a7484405/sensors-22-05910-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/14b56752dcbe/sensors-22-05910-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/38c7b1ed2508/sensors-22-05910-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/ca6b0b376db5/sensors-22-05910-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/b3a5a6c96ae0/sensors-22-05910-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/31a53b8714ef/sensors-22-05910-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/638869a1fb7b/sensors-22-05910-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/850ea1987b4e/sensors-22-05910-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/b02b08dbc029/sensors-22-05910-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/ff29a3f53e35/sensors-22-05910-g017.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/dfba416e1e82/sensors-22-05910-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/1e639d3e42da/sensors-22-05910-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/0ca40cb3cba5/sensors-22-05910-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/455f67c87226/sensors-22-05910-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/bfc6053989b3/sensors-22-05910-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/533524974360/sensors-22-05910-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/8119fe092d5f/sensors-22-05910-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/47d6a7484405/sensors-22-05910-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/14b56752dcbe/sensors-22-05910-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/38c7b1ed2508/sensors-22-05910-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/ca6b0b376db5/sensors-22-05910-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/b3a5a6c96ae0/sensors-22-05910-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/31a53b8714ef/sensors-22-05910-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/638869a1fb7b/sensors-22-05910-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/850ea1987b4e/sensors-22-05910-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/b02b08dbc029/sensors-22-05910-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d303/9371426/ff29a3f53e35/sensors-22-05910-g017.jpg

相似文献

1
CLSQL: Improved Q-Learning Algorithm Based on Continuous Local Search Policy for Mobile Robot Path Planning.CLSQL:基于连续局部搜索策略的移动机器人路径规划改进Q学习算法
Sensors (Basel). 2022 Aug 8;22(15):5910. doi: 10.3390/s22155910.
2
A path planning approach for mobile robots using short and safe Q-learning.基于短程且安全 Q-学习的移动机器人路径规划方法。
PLoS One. 2022 Sep 26;17(9):e0275100. doi: 10.1371/journal.pone.0275100. eCollection 2022.
3
A Path-Planning Approach Based on Potential and Dynamic Q-Learning for Mobile Robots in Unknown Environment.基于势场和动态 Q 学习的未知环境下移动机器人路径规划方法。
Comput Intell Neurosci. 2022 Jun 2;2022:2540546. doi: 10.1155/2022/2540546. eCollection 2022.
4
The Path Planning of Mobile Robot by Neural Networks and Hierarchical Reinforcement Learning.基于神经网络和分层强化学习的移动机器人路径规划
Front Neurorobot. 2020 Oct 2;14:63. doi: 10.3389/fnbot.2020.00063. eCollection 2020.
5
Efficient Path Planning for Mobile Robot Based on Deep Deterministic Policy Gradient.基于深度确定性策略梯度的移动机器人高效路径规划。
Sensors (Basel). 2022 May 8;22(9):3579. doi: 10.3390/s22093579.
6
LF-ACO: an effective formation path planning for multi-mobile robot.LF-ACO:一种用于多移动机器人的有效编队路径规划方法。
Math Biosci Eng. 2022 Jan;19(1):225-252. doi: 10.3934/mbe.2022012. Epub 2021 Nov 9.
7
A search and rescue robot search method based on flower pollination algorithm and Q-learning fusion algorithm.基于花授粉算法和 Q 学习融合算法的搜索救援机器人搜索方法。
PLoS One. 2023 Mar 30;18(3):e0283751. doi: 10.1371/journal.pone.0283751. eCollection 2023.
8
Particle Swarm Algorithm Path-Planning Method for Mobile Robots Based on Artificial Potential Fields.基于人工势场的移动机器人粒子群算法路径规划方法。
Sensors (Basel). 2023 Jul 1;23(13):6082. doi: 10.3390/s23136082.
9
Path Planning for Wheeled Mobile Robot in Partially Known Uneven Terrain.轮式移动机器人在部分未知非均匀地形中的路径规划。
Sensors (Basel). 2022 Jul 12;22(14):5217. doi: 10.3390/s22145217.
10
Intelligent Optimization Algorithm-Based Path Planning for a Mobile Robot.基于智能优化算法的移动机器人路径规划。
Comput Intell Neurosci. 2021 Sep 29;2021:8025730. doi: 10.1155/2021/8025730. eCollection 2021.

引用本文的文献

1
Path Planning Trends for Autonomous Mobile Robot Navigation: A Review.自主移动机器人导航的路径规划趋势:综述
Sensors (Basel). 2025 Feb 16;25(4):1206. doi: 10.3390/s25041206.

本文引用的文献

1
Path Planning for Wheeled Mobile Robot in Partially Known Uneven Terrain.轮式移动机器人在部分未知非均匀地形中的路径规划。
Sensors (Basel). 2022 Jul 12;22(14):5217. doi: 10.3390/s22145217.
2
A Hierarchical Path Planning Approach with Multi-SARSA Based on Topological Map.一种基于拓扑地图的多SARSA分层路径规划方法。
Sensors (Basel). 2022 Mar 18;22(6):2367. doi: 10.3390/s22062367.
3
Reinforcement learning of motor skills with policy gradients.基于策略梯度的运动技能强化学习。
Neural Netw. 2008 May;21(4):682-97. doi: 10.1016/j.neunet.2008.02.003. Epub 2008 Apr 26.