• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

具有水质传感器的自主水面车辆的进化算法与深度强化学习方法的维度比较。

A Dimensional Comparison between Evolutionary Algorithm and Deep Reinforcement Learning Methodologies for Autonomous Surface Vehicles with Water Quality Sensors.

机构信息

Department of Electronic Engineering, University of Seville, 41009 Seville, Spain.

出版信息

Sensors (Basel). 2021 Apr 19;21(8):2862. doi: 10.3390/s21082862.

DOI:10.3390/s21082862
PMID:33921649
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8074202/
Abstract

The monitoring of water resources using Autonomous Surface Vehicles with water-quality sensors has been a recent approach due to the advances in unmanned transportation technology. The Ypacaraí Lake, the biggest water resource in Paraguay, suffers from a major contamination problem because of cyanobacteria blooms. In order to supervise the blooms using these on-board sensor modules, a Non-Homogeneous Patrolling Problem (a NP-hard problem) must be solved in a feasible amount of time. A dimensionality study is addressed to compare the most common methodologies, Evolutionary Algorithm and Deep Reinforcement Learning, in different map scales and fleet sizes with changes in the environmental conditions. The results determined that Deep Q-Learning overcomes the evolutionary method in terms of sample-efficiency by 50-70% in higher resolutions. Furthermore, it reacts better than the Evolutionary Algorithm in high space-state actions. In contrast, the evolutionary approach shows a better efficiency in lower resolutions and needs fewer parameters to synthesize robust solutions. This study reveals that Deep Q-learning approaches exceed in efficiency for the Non-Homogeneous Patrolling Problem but with many hyper-parameters involved in the stability and convergence.

摘要

利用带有水质传感器的自主水面车辆监测水资源是最近的一种方法,因为无人运输技术取得了进步。亚帕克拉伊湖是巴拉圭最大的水资源,但由于蓝藻水华,它面临着严重的污染问题。为了使用这些机载传感器模块来监测水华,必须在可行的时间内解决非均匀巡逻问题(一个 NP 难问题)。本研究进行了维度研究,以比较最常见的方法,进化算法和深度强化学习,在不同的地图比例和船队规模与环境条件的变化。结果表明,深度 Q 学习在样本效率方面比进化方法提高了 50-70%,在更高的分辨率下。此外,它在高空间状态动作中的反应优于进化算法。相比之下,进化方法在较低的分辨率下显示出更好的效率,并且需要较少的参数来综合鲁棒的解决方案。这项研究表明,深度 Q 学习方法在非均匀巡逻问题上的效率更高,但在稳定性和收敛性方面涉及到许多超参数。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/97226472ee45/sensors-21-02862-g020.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/189771ca3688/sensors-21-02862-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/326b58fc79ee/sensors-21-02862-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/e067824b1696/sensors-21-02862-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/c62e66e9f301/sensors-21-02862-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/4ca8c3f9c64a/sensors-21-02862-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/c594618827f2/sensors-21-02862-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/03571cbdc345/sensors-21-02862-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/3a3a9946a014/sensors-21-02862-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/a38723a22423/sensors-21-02862-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/45a641940946/sensors-21-02862-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/f335a4053abd/sensors-21-02862-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/b98a03f54ec2/sensors-21-02862-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/702096a75cf8/sensors-21-02862-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/5ac3bffd58b3/sensors-21-02862-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/b2bb85e60328/sensors-21-02862-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/a858d86b649e/sensors-21-02862-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/a4c33b090c3a/sensors-21-02862-g017.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/fabe69b920a0/sensors-21-02862-g018.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/d97e141e05a7/sensors-21-02862-g019.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/97226472ee45/sensors-21-02862-g020.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/189771ca3688/sensors-21-02862-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/326b58fc79ee/sensors-21-02862-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/e067824b1696/sensors-21-02862-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/c62e66e9f301/sensors-21-02862-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/4ca8c3f9c64a/sensors-21-02862-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/c594618827f2/sensors-21-02862-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/03571cbdc345/sensors-21-02862-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/3a3a9946a014/sensors-21-02862-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/a38723a22423/sensors-21-02862-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/45a641940946/sensors-21-02862-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/f335a4053abd/sensors-21-02862-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/b98a03f54ec2/sensors-21-02862-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/702096a75cf8/sensors-21-02862-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/5ac3bffd58b3/sensors-21-02862-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/b2bb85e60328/sensors-21-02862-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/a858d86b649e/sensors-21-02862-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/a4c33b090c3a/sensors-21-02862-g017.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/fabe69b920a0/sensors-21-02862-g018.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/d97e141e05a7/sensors-21-02862-g019.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3085/8074202/97226472ee45/sensors-21-02862-g020.jpg

相似文献

1
A Dimensional Comparison between Evolutionary Algorithm and Deep Reinforcement Learning Methodologies for Autonomous Surface Vehicles with Water Quality Sensors.具有水质传感器的自主水面车辆的进化算法与深度强化学习方法的维度比较。
Sensors (Basel). 2021 Apr 19;21(8):2862. doi: 10.3390/s21082862.
2
Intelligent Vehicle Decision-Making and Trajectory Planning Method Based on Deep Reinforcement Learning in the Frenet Space.基于弗伦内特空间深度强化学习的智能车辆决策与轨迹规划方法
Sensors (Basel). 2023 Dec 14;23(24):9819. doi: 10.3390/s23249819.
3
UAV Autonomous Tracking and Landing Based on Deep Reinforcement Learning Strategy.基于深度强化学习策略的无人机自主跟踪与着陆
Sensors (Basel). 2020 Oct 1;20(19):5630. doi: 10.3390/s20195630.
4
Decision-Making for the Autonomous Navigation of Maritime Autonomous Surface Ships Based on Scene Division and Deep Reinforcement Learning.基于场景划分和深度强化学习的航海自主水面船舶自主导航决策。
Sensors (Basel). 2019 Sep 19;19(18):4055. doi: 10.3390/s19184055.
5
A Turbo Q-Learning (TQL) for Energy Efficiency Optimization in Heterogeneous Networks.一种用于异构网络能效优化的Turbo Q学习(TQL)
Entropy (Basel). 2020 Aug 30;22(9):957. doi: 10.3390/e22090957.
6
Double Deep Q-Learning and Faster R-CNN-Based Autonomous Vehicle Navigation and Obstacle Avoidance in Dynamic Environment.基于双深度 Q 学习和 Faster R-CNN 的动态环境下自主车辆导航和避障。
Sensors (Basel). 2021 Feb 20;21(4):1468. doi: 10.3390/s21041468.
7
Robot grasping method optimization using improved deep deterministic policy gradient algorithm of deep reinforcement learning.基于深度强化学习的改进深度确定性策略梯度算法的机器人抓取方法优化
Rev Sci Instrum. 2021 Feb 1;92(2):025114. doi: 10.1063/5.0034101.
8
An Autonomous Path Planning Model for Unmanned Ships Based on Deep Reinforcement Learning.基于深度强化学习的无人船自主路径规划模型。
Sensors (Basel). 2020 Jan 11;20(2):426. doi: 10.3390/s20020426.
9
Deep Reinforcement Learning Approach with Multiple Experience Pools for UAV's Autonomous Motion Planning in Complex Unknown Environments.深度强化学习方法与多经验池在复杂未知环境中无人机自主运动规划。
Sensors (Basel). 2020 Mar 29;20(7):1890. doi: 10.3390/s20071890.
10
Domain Knowledge-Based Evolutionary Reinforcement Learning for Sensor Placement.基于领域知识的进化强化学习在传感器布放中的应用。
Sensors (Basel). 2022 May 17;22(10):3799. doi: 10.3390/s22103799.

引用本文的文献

1
Heuristic Greedy Scheduling of Electric Vehicles in Vehicle-to-Grid Microgrid Owned Aggregators.车网互动微电网拥有者的启发式贪婪电动汽车调度。
Sensors (Basel). 2022 Mar 21;22(6):2408. doi: 10.3390/s22062408.

本文引用的文献

1
Autonomous Water Quality Monitoring and Water Surface Cleaning for Unmanned Surface Vehicle.自主水质监测和水面清洁的无人水面艇。
Sensors (Basel). 2021 Feb 5;21(4):1102. doi: 10.3390/s21041102.
2
Reinforcement Learning-Based Complete Area Coverage Path Planning for a Modified hTrihex Robot.基于强化学习的改进 hTrihex 机器人全覆盖路径规划
Sensors (Basel). 2021 Feb 4;21(4):1067. doi: 10.3390/s21041067.
3
A Comparison of Local Path Planning Techniques of Autonomous Surface Vehicles for Monitoring Applications: The Ypacarai Lake Case-study.
自主水面船舶用于监测应用的局部路径规划技术比较:亚帕克瑞亚湖案例研究。
Sensors (Basel). 2020 Mar 9;20(5):1488. doi: 10.3390/s20051488.
4
Modular AUV System with Integrated Real-Time Water Quality Analysis.模块化自主水下航行器系统,集成实时水质分析功能。
Sensors (Basel). 2018 Jun 5;18(6):1837. doi: 10.3390/s18061837.
5
Human-level control through deep reinforcement learning.通过深度强化学习实现人类水平的控制。
Nature. 2015 Feb 26;518(7540):529-33. doi: 10.1038/nature14236.