• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

通过地面到水面跨域深度强化学习实现稳健的自主水面航行器导航

Robust ASV Navigation Through Ground to Water Cross-Domain Deep Reinforcement Learning.

作者信息

Lambert Reeve, Li Jianwen, Wu Li-Fan, Mahmoudian Nina

机构信息

MS Student, School of Mechanical Engineering, Purdue University, West Lafayette, IN, United States.

PhD Student, School of Mechanical Engineering, Purdue University, West Lafayette, IN, United States.

出版信息

Front Robot AI. 2021 Sep 20;8:739023. doi: 10.3389/frobt.2021.739023. eCollection 2021.

DOI:10.3389/frobt.2021.739023
PMID:34616776
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8488128/
Abstract

This paper presents a framework to alleviate the Deep Reinforcement Learning (DRL) training data sparsity problem that is present in challenging domains by creating a DRL agent training and vehicle integration methodology. The methodology leverages accessible domains to train an agent to solve navigational problems such as obstacle avoidance and allows the agent to generalize to challenging and inaccessible domains such as those present in marine environments with minimal further training. This is done by integrating a DRL agent at a high level of vehicle control and leveraging existing path planning and proven low-level control methodologies that are utilized in multiple domains. An autonomy package with a tertiary multilevel controller is developed to enable the DRL agent to interface at the prescribed high control level and thus be separated from vehicle dynamics and environmental constraints. An example Deep Q Network (DQN) employing this methodology for obstacle avoidance is trained in a simulated ground environment, and then its ability to generalize across domains is experimentally validated. Experimental validation utilized a simulated water surface environment and real-world deployment of ground and water robotic platforms. This methodology, when used, shows that it is possible to leverage accessible and data rich domains, such as ground, to effectively develop marine DRL agents for use on Autonomous Surface Vehicle (ASV) navigation. This will allow rapid and iterative agent development without the risk of ASV loss, the cost and logistic overhead of marine deployment, and allow landlocked institutions to develop agents for marine applications.

摘要

本文提出了一个框架,通过创建一种深度强化学习(DRL)智能体训练与车辆集成方法,来缓解在具有挑战性的领域中存在的DRL训练数据稀疏问题。该方法利用可访问的领域来训练智能体解决诸如避障等导航问题,并使智能体在经过最少的进一步训练后,能够推广到具有挑战性和不可访问的领域,如海洋环境中存在的领域。这是通过在车辆控制的高级层面集成DRL智能体,并利用现有的路径规划和经过验证的低级控制方法来实现的,这些方法在多个领域中都有应用。开发了一个带有三级多级控制器的自主软件包,以使DRL智能体能够在规定的高控制级别进行交互,从而与车辆动力学和环境约束分离。一个采用这种避障方法的深度Q网络(DQN)示例在模拟地面环境中进行了训练,然后通过实验验证了其跨领域推广的能力。实验验证利用了模拟水面环境以及地面和水上机器人平台的实际部署。这种方法在使用时表明,利用诸如地面等可访问且数据丰富的领域,有效地开发用于自主水面舰艇(ASV)导航的海洋DRL智能体是可能的。这将允许快速且迭代地开发智能体,而不会有ASV损失的风险、海洋部署的成本和后勤开销,并且允许内陆机构开发用于海洋应用的智能体。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7622/8488128/97ba3d6b4567/frobt-08-739023-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7622/8488128/7109ff7812f1/frobt-08-739023-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7622/8488128/344ce6d51760/frobt-08-739023-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7622/8488128/a001cad0c502/frobt-08-739023-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7622/8488128/439854ca35fa/frobt-08-739023-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7622/8488128/da236eaab563/frobt-08-739023-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7622/8488128/b8e0264dcc71/frobt-08-739023-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7622/8488128/afc404d347b7/frobt-08-739023-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7622/8488128/71a60242215c/frobt-08-739023-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7622/8488128/9f2ab06a52a5/frobt-08-739023-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7622/8488128/97ba3d6b4567/frobt-08-739023-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7622/8488128/7109ff7812f1/frobt-08-739023-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7622/8488128/344ce6d51760/frobt-08-739023-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7622/8488128/a001cad0c502/frobt-08-739023-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7622/8488128/439854ca35fa/frobt-08-739023-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7622/8488128/da236eaab563/frobt-08-739023-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7622/8488128/b8e0264dcc71/frobt-08-739023-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7622/8488128/afc404d347b7/frobt-08-739023-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7622/8488128/71a60242215c/frobt-08-739023-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7622/8488128/9f2ab06a52a5/frobt-08-739023-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7622/8488128/97ba3d6b4567/frobt-08-739023-g010.jpg

相似文献

1
Robust ASV Navigation Through Ground to Water Cross-Domain Deep Reinforcement Learning.通过地面到水面跨域深度强化学习实现稳健的自主水面航行器导航
Front Robot AI. 2021 Sep 20;8:739023. doi: 10.3389/frobt.2021.739023. eCollection 2021.
2
Dynamic Obstacle Avoidance for USVs Using Cross-Domain Deep Reinforcement Learning and Neural Network Model Predictive Controller.基于跨域深度强化学习和神经网络模型预测控制器的 USV 动态避障
Sensors (Basel). 2023 Mar 29;23(7):3572. doi: 10.3390/s23073572.
3
Decision-Making for the Autonomous Navigation of Maritime Autonomous Surface Ships Based on Scene Division and Deep Reinforcement Learning.基于场景划分和深度强化学习的航海自主水面船舶自主导航决策。
Sensors (Basel). 2019 Sep 19;19(18):4055. doi: 10.3390/s19184055.
4
Double Deep Q-Learning and Faster R-CNN-Based Autonomous Vehicle Navigation and Obstacle Avoidance in Dynamic Environment.基于双深度 Q 学习和 Faster R-CNN 的动态环境下自主车辆导航和避障。
Sensors (Basel). 2021 Feb 20;21(4):1468. doi: 10.3390/s21041468.
5
Risk-based implementation of COLREGs for autonomous surface vehicles using deep reinforcement learning.基于风险的自主水面船舶 COLREGs 实施的深度学习强化学习方法。
Neural Netw. 2022 Aug;152:17-33. doi: 10.1016/j.neunet.2022.04.008. Epub 2022 Apr 16.
6
Deep reinforcement learning for automated radiation adaptation in lung cancer.深度强化学习在肺癌放射自适应中的应用。
Med Phys. 2017 Dec;44(12):6690-6705. doi: 10.1002/mp.12625. Epub 2017 Nov 14.
7
Predictive hierarchical reinforcement learning for path-efficient mapless navigation with moving target.具有移动目标的无图路径高效导航的预测分层强化学习。
Neural Netw. 2023 Aug;165:677-688. doi: 10.1016/j.neunet.2023.06.007. Epub 2023 Jun 10.
8
Mobile Robot Application with Hierarchical Start Position DQN.分层起始位置 DQN 的移动机器人应用。
Comput Intell Neurosci. 2022 Sep 5;2022:4115767. doi: 10.1155/2022/4115767. eCollection 2022.
9
Deep reinforcement learning-aided autonomous navigation with landmark generators.基于地标生成器的深度强化学习辅助自主导航。
Front Neurorobot. 2023 Aug 22;17:1200214. doi: 10.3389/fnbot.2023.1200214. eCollection 2023.
10
Improved Robot Path Planning Method Based on Deep Reinforcement Learning.基于深度强化学习的改进型机器人路径规划方法。
Sensors (Basel). 2023 Jun 15;23(12):5622. doi: 10.3390/s23125622.

引用本文的文献

1
Dynamic Obstacle Avoidance for USVs Using Cross-Domain Deep Reinforcement Learning and Neural Network Model Predictive Controller.基于跨域深度强化学习和神经网络模型预测控制器的 USV 动态避障
Sensors (Basel). 2023 Mar 29;23(7):3572. doi: 10.3390/s23073572.

本文引用的文献

1
A general reinforcement learning algorithm that masters chess, shogi, and Go through self-play.一种通过自我对弈掌握国际象棋、将棋和围棋的通用强化学习算法。
Science. 2018 Dec 7;362(6419):1140-1144. doi: 10.1126/science.aar6404.
2
Rapid Mapping of Dissolved Methane and Carbon Dioxide in Coastal Ecosystems Using the ChemYak Autonomous Surface Vehicle.利用 ChemYak 自主水面艇快速绘制沿海生态系统中溶解甲烷和二氧化碳的分布图。
Environ Sci Technol. 2018 Nov 20;52(22):13314-13324. doi: 10.1021/acs.est.8b04190. Epub 2018 Nov 7.
3
Human-level control through deep reinforcement learning.
通过深度强化学习实现人类水平的控制。
Nature. 2015 Feb 26;518(7540):529-33. doi: 10.1038/nature14236.