• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

分层起始位置 DQN 的移动机器人应用。

Mobile Robot Application with Hierarchical Start Position DQN.

机构信息

Department of Electronic Communication, Batman University, Batman 72500, Turkey.

Electrical and Electronics Engineering Department, Dicle University, Diyarbakir 21280, Turkey.

出版信息

Comput Intell Neurosci. 2022 Sep 5;2022:4115767. doi: 10.1155/2022/4115767. eCollection 2022.

DOI:10.1155/2022/4115767
PMID:36105641
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC9467786/
Abstract

Advances in deep learning significantly affect reinforcement learning, which results in the emergence of Deep RL (DRL). DRL does not need a data set and has the potential beyond the performance of human experts, resulting in significant developments in the field of artificial intelligence. However, because a DRL agent has to interact with the environment a lot while it is trained, it is difficult to be trained directly in the real environment due to the long training time, high cost, and possible material damage. Therefore, most or all of the training of DRL agents for real-world applications is conducted in virtual environments. This study focused on the difficulty in a mobile robot to reach its target by making a path plan in a real-world environment. The Minimalistic Gridworld virtual environment has been used for training the DRL agent, and to our knowledge, we have implemented the first real-world implementation for this environment. A DRL algorithm with higher performance than the classical Deep Q-network algorithm was created with the expanded environment. A mobile robot was designed for use in a real-world application. To match the virtual environment with the real environment, algorithms that can detect the position of the mobile robot and the target, as well as the rotation of the mobile robot, were created. As a result, a DRL-based mobile robot was developed that uses only the top view of the environment and can reach its target regardless of its initial position and rotation.

摘要

深度学习的进展极大地影响了强化学习,从而产生了深度强化学习(DRL)。DRL 不需要数据集,并且具有超越人类专家表现的潜力,从而在人工智能领域取得了重大进展。然而,由于 DRL 代理在训练过程中必须与环境进行大量交互,因此由于训练时间长、成本高和可能的材料损坏,直接在真实环境中进行训练是困难的。因此,对于现实世界应用的 DRL 代理的大部分或全部训练都是在虚拟环境中进行的。本研究侧重于移动机器人在真实环境中通过制定路径计划到达目标时所面临的困难。最小网格世界虚拟环境已被用于训练 DRL 代理,据我们所知,我们已经为该环境实现了第一个真实世界的实现。使用扩展环境创建了一种性能优于经典深度 Q 网络算法的 DRL 算法。设计了一种移动机器人用于现实世界的应用。为了使虚拟环境与真实环境相匹配,创建了可以检测移动机器人和目标的位置以及移动机器人的旋转的算法。结果,开发了一种基于 DRL 的移动机器人,它仅使用环境的俯视图,并且可以到达其目标,而不管其初始位置和旋转如何。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fc25/9467786/79db2d27c4f7/CIN2022-4115767.002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fc25/9467786/79db2d27c4f7/CIN2022-4115767.002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fc25/9467786/79db2d27c4f7/CIN2022-4115767.002.jpg

相似文献

1
Mobile Robot Application with Hierarchical Start Position DQN.分层起始位置 DQN 的移动机器人应用。
Comput Intell Neurosci. 2022 Sep 5;2022:4115767. doi: 10.1155/2022/4115767. eCollection 2022.
2
Improved Robot Path Planning Method Based on Deep Reinforcement Learning.基于深度强化学习的改进型机器人路径规划方法。
Sensors (Basel). 2023 Jun 15;23(12):5622. doi: 10.3390/s23125622.
3
Deep Reinforcement Learning for Indoor Mobile Robot Path Planning.深度强化学习在室内移动机器人路径规划中的应用。
Sensors (Basel). 2020 Sep 25;20(19):5493. doi: 10.3390/s20195493.
4
Predictive hierarchical reinforcement learning for path-efficient mapless navigation with moving target.具有移动目标的无图路径高效导航的预测分层强化学习。
Neural Netw. 2023 Aug;165:677-688. doi: 10.1016/j.neunet.2023.06.007. Epub 2023 Jun 10.
5
SLAM algorithm applied to robotics assistance for navigation in unknown environments.SLAM 算法在机器人辅助未知环境导航中的应用。
J Neuroeng Rehabil. 2010 Feb 17;7:10. doi: 10.1186/1743-0003-7-10.
6
Approximate Policy-Based Accelerated Deep Reinforcement Learning.基于近似策略的加速深度强化学习
IEEE Trans Neural Netw Learn Syst. 2020 Jun;31(6):1820-1830. doi: 10.1109/TNNLS.2019.2927227. Epub 2019 Aug 6.
7
Multimodal Deep Reinforcement Learning with Auxiliary Task for Obstacle Avoidance of Indoor Mobile Robot.多模态深度强化学习与辅助任务在室内移动机器人避障中的应用。
Sensors (Basel). 2021 Feb 15;21(4):1363. doi: 10.3390/s21041363.
8
Deep reinforcement learning for automated radiation adaptation in lung cancer.深度强化学习在肺癌放射自适应中的应用。
Med Phys. 2017 Dec;44(12):6690-6705. doi: 10.1002/mp.12625. Epub 2017 Nov 14.
9
Deep reinforcement learning and its applications in medical imaging and radiation therapy: a survey.深度强化学习及其在医学影像和放射治疗中的应用:综述。
Phys Med Biol. 2022 Nov 11;67(22). doi: 10.1088/1361-6560/ac9cb3.
10
Dynamic Obstacle Avoidance for USVs Using Cross-Domain Deep Reinforcement Learning and Neural Network Model Predictive Controller.基于跨域深度强化学习和神经网络模型预测控制器的 USV 动态避障
Sensors (Basel). 2023 Mar 29;23(7):3572. doi: 10.3390/s23073572.

本文引用的文献

1
Efficient Path Planning for Mobile Robot Based on Deep Deterministic Policy Gradient.基于深度确定性策略梯度的移动机器人高效路径规划。
Sensors (Basel). 2022 May 8;22(9):3579. doi: 10.3390/s22093579.
2
Modular deep reinforcement learning from reward and punishment for robot navigation.基于奖惩的机器人导航模块化深度强化学习。
Neural Netw. 2021 Mar;135:115-126. doi: 10.1016/j.neunet.2020.12.001. Epub 2020 Dec 8.
3
The Path Planning of Mobile Robot by Neural Networks and Hierarchical Reinforcement Learning.基于神经网络和分层强化学习的移动机器人路径规划
Front Neurorobot. 2020 Oct 2;14:63. doi: 10.3389/fnbot.2020.00063. eCollection 2020.
4
Intrinsically motivated reinforcement learning for human-robot interaction in the real-world.基于内在动机的强化学习在真实世界中的人机交互
Neural Netw. 2018 Nov;107:23-33. doi: 10.1016/j.neunet.2018.03.014. Epub 2018 Mar 26.
5
Mastering the game of Go without human knowledge.无需人类知识即可掌握围棋游戏。
Nature. 2017 Oct 18;550(7676):354-359. doi: 10.1038/nature24270.
6
Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks.更快的 R-CNN:基于区域建议网络的实时目标检测。
IEEE Trans Pattern Anal Mach Intell. 2017 Jun;39(6):1137-1149. doi: 10.1109/TPAMI.2016.2577031. Epub 2016 Jun 6.
7
Human-level control through deep reinforcement learning.通过深度强化学习实现人类水平的控制。
Nature. 2015 Feb 26;518(7540):529-33. doi: 10.1038/nature14236.
8
On the origin of the bilateral filter and ways to improve it.关于双边滤波器的起源及改进方法。
IEEE Trans Image Process. 2002;11(10):1141-51. doi: 10.1109/TIP.2002.801126.