• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于强化学习的无人水面艇在不同运行条件下的跟踪控制

Reinforcement Learning-Based Tracking Control of USVs in Varying Operational Conditions.

作者信息

Martinsen Andreas B, Lekkas Anastasios M, Gros Sébastien, Glomsrud Jon Arne, Pedersen Tom Arne

机构信息

Department of Engineering Cybernetics, Norwegian University of Science and Technology, Trondheim, Norway.

Centre for Autonomous Marine Operations and Systems, Norwegian University of Science and Technology, Trondheim, Norway.

出版信息

Front Robot AI. 2020 Mar 20;7:32. doi: 10.3389/frobt.2020.00032. eCollection 2020.

DOI:10.3389/frobt.2020.00032
PMID:33501200
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7806118/
Abstract

We present a reinforcement learning-based (RL) control scheme for trajectory tracking of fully-actuated surface vessels. The proposed method learns online both a model-based feedforward controller, as well an optimizing feedback policy in order to follow a desired trajectory under the influence of environmental forces. The method's efficiency is evaluated via simulations and sea trials, with the unmanned surface vehicle (USV) performing three different tracking tasks: The four corner DP test, straight-path tracking and curved-path tracking. The results demonstrate the method's ability to accomplish the control objectives and a good agreement between the performance achieved in the Revolt Digital Twin and the sea trials. Finally, we include an section with considerations about assurance for RL-based methods and where our approach stands in terms of the main challenges.

摘要

我们提出了一种基于强化学习(RL)的控制方案,用于全驱动水面舰艇的轨迹跟踪。所提出的方法在线学习基于模型的前馈控制器以及优化反馈策略,以便在环境力的影响下跟踪期望轨迹。通过仿真和海上试验评估了该方法的效率,无人水面舰艇(USV)执行了三种不同的跟踪任务:四角动力定位测试、直线路径跟踪和曲线路径跟踪。结果证明了该方法实现控制目标的能力,以及在Revolt数字孪生中实现的性能与海上试验之间的良好一致性。最后,我们纳入了一个部分,其中考虑了基于RL方法的保证,以及我们的方法在主要挑战方面的立场。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e68c/7806118/0805a0a3b9e6/frobt-07-00032-g0011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e68c/7806118/b033e7c6d749/frobt-07-00032-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e68c/7806118/f4e4a41abe02/frobt-07-00032-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e68c/7806118/389d2af7ce15/frobt-07-00032-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e68c/7806118/aef8fe601f7a/frobt-07-00032-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e68c/7806118/830c4c8f5118/frobt-07-00032-g0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e68c/7806118/9ab50f5ef932/frobt-07-00032-g0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e68c/7806118/c087fafdfedd/frobt-07-00032-g0007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e68c/7806118/fc29b0573ba1/frobt-07-00032-g0008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e68c/7806118/0c257a50d25e/frobt-07-00032-g0009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e68c/7806118/2b390e3cf28b/frobt-07-00032-g0010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e68c/7806118/0805a0a3b9e6/frobt-07-00032-g0011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e68c/7806118/b033e7c6d749/frobt-07-00032-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e68c/7806118/f4e4a41abe02/frobt-07-00032-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e68c/7806118/389d2af7ce15/frobt-07-00032-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e68c/7806118/aef8fe601f7a/frobt-07-00032-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e68c/7806118/830c4c8f5118/frobt-07-00032-g0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e68c/7806118/9ab50f5ef932/frobt-07-00032-g0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e68c/7806118/c087fafdfedd/frobt-07-00032-g0007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e68c/7806118/fc29b0573ba1/frobt-07-00032-g0008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e68c/7806118/0c257a50d25e/frobt-07-00032-g0009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e68c/7806118/2b390e3cf28b/frobt-07-00032-g0010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e68c/7806118/0805a0a3b9e6/frobt-07-00032-g0011.jpg

相似文献

1
Reinforcement Learning-Based Tracking Control of USVs in Varying Operational Conditions.基于强化学习的无人水面艇在不同运行条件下的跟踪控制
Front Robot AI. 2020 Mar 20;7:32. doi: 10.3389/frobt.2020.00032. eCollection 2020.
2
Path Following Control for Unmanned Surface Vehicles: A Reinforcement Learning-Based Method With Experimental Validation.无人水面舰艇的路径跟踪控制:一种基于强化学习的方法及实验验证
IEEE Trans Neural Netw Learn Syst. 2024 Dec;35(12):18237-18250. doi: 10.1109/TNNLS.2023.3313312. Epub 2024 Dec 2.
3
USV Trajectory Tracking Control Based on Receding Horizon Reinforcement Learning.基于滚动时域强化学习的无人水面艇轨迹跟踪控制
Sensors (Basel). 2024 Apr 26;24(9):2771. doi: 10.3390/s24092771.
4
Adaptive Optimal Surrounding Control of Multiple Unmanned Surface Vessels via Actor-Critic Reinforcement Learning.基于智能体-评论家强化学习的多艘无人水面舰艇自适应最优周边控制
IEEE Trans Neural Netw Learn Syst. 2024 Oct 18;PP. doi: 10.1109/TNNLS.2024.3474289.
5
Adaptive optimal trajectory tracking control of AUVs based on reinforcement learning.基于强化学习的 AUV 自适应最优轨迹跟踪控制。
ISA Trans. 2023 Jun;137:122-132. doi: 10.1016/j.isatra.2022.12.003. Epub 2022 Dec 8.
6
Multi-under-Actuated Unmanned Surface Vessel Coordinated Path Tracking.多欠驱动无人水面艇协同路径跟踪
Sensors (Basel). 2020 Feb 6;20(3):864. doi: 10.3390/s20030864.
7
Dynamic Navigation and Area Assignment of Multiple USVs Based on Multi-Agent Deep Reinforcement Learning.基于多智能体深度强化学习的多艘 USV 的动态导航与区域分配。
Sensors (Basel). 2022 Sep 14;22(18):6942. doi: 10.3390/s22186942.
8
Reinforcement Learning Tracking Control for Robotic Manipulator With Kernel-Based Dynamic Model.基于核动态模型的机器人机械手强化学习跟踪控制
IEEE Trans Neural Netw Learn Syst. 2020 Sep;31(9):3570-3578. doi: 10.1109/TNNLS.2019.2945019. Epub 2019 Nov 1.
9
Time-varying gain extended state observer-based adaptive optimal control for disturbed unmanned helicopter.基于时变增益扩张状态观测器的干扰无人直升机自适应最优控制
ISA Trans. 2024 May;148:1-11. doi: 10.1016/j.isatra.2024.02.028. Epub 2024 Feb 28.
10
Data-Driven Performance-Prescribed Reinforcement Learning Control of an Unmanned Surface Vehicle.基于数据驱动的无人水面艇性能规定强化学习控制
IEEE Trans Neural Netw Learn Syst. 2021 Dec;32(12):5456-5467. doi: 10.1109/TNNLS.2021.3056444. Epub 2021 Nov 30.

引用本文的文献

1
Fuzzy adaptive fault-tolerant control for an unmanned surface vehicle with prescribed tracking performance.具有规定跟踪性能的无人水面艇模糊自适应容错控制
Front Robot AI. 2025 Apr 9;12:1576171. doi: 10.3389/frobt.2025.1576171. eCollection 2025.
2
Torque coordinated control of the through-the-road (TTR) 4-wheel-drive (4WD) hybrid vehicle under extreme road conditions.极端道路条件下贯通式四轮驱动(TTR)混合动力车辆的扭矩协调控制
Sci Rep. 2023 Jul 18;13(1):11564. doi: 10.1038/s41598-023-38813-3.
3
Estimating spatio-temporal fields through reinforcement learning.

本文引用的文献

1
A general reinforcement learning algorithm that masters chess, shogi, and Go through self-play.一种通过自我对弈掌握国际象棋、将棋和围棋的通用强化学习算法。
Science. 2018 Dec 7;362(6419):1140-1144. doi: 10.1126/science.aar6404.
2
Mastering the game of Go with deep neural networks and tree search.用深度神经网络和树搜索掌握围棋游戏。
Nature. 2016 Jan 28;529(7587):484-9. doi: 10.1038/nature16961.
3
Reinforcement learning in continuous time and space.连续时间与空间中的强化学习。
通过强化学习估计时空场。
Front Robot AI. 2022 Sep 5;9:878246. doi: 10.3389/frobt.2022.878246. eCollection 2022.
Neural Comput. 2000 Jan;12(1):219-45. doi: 10.1162/089976600300015961.