• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于混沌时空动力学深度强化学习主动控制的对称性约化

Symmetry reduction for deep reinforcement learning active control of chaotic spatiotemporal dynamics.

作者信息

Zeng Kevin, Graham Michael D

机构信息

Department of Chemical and Biological Engineering, University of Wisconsin-Madison, Madison, Wisconsin 53706, USA.

出版信息

Phys Rev E. 2021 Jul;104(1-1):014210. doi: 10.1103/PhysRevE.104.014210.

DOI:10.1103/PhysRevE.104.014210
PMID:34412246
Abstract

Deep reinforcement learning (RL) is a data-driven, model-free method capable of discovering complex control strategies for macroscopic objectives in high-dimensional systems, making its application toward flow control promising. Many systems of flow control interest possess symmetries that, when neglected, can significantly inhibit the learning and performance of a naive deep RL approach. Using a test-bed consisting of the Kuramoto-Sivashinsky equation (KSE), equally spaced actuators, and a goal of minimizing dissipation and power cost, we demonstrate that by moving the deep RL problem to a symmetry-reduced space, we can alleviate limitations inherent in the naive application of deep RL. We demonstrate that symmetry-reduced deep RL yields improved data efficiency as well as improved control policy efficacy compared to policies found by naive deep RL. Interestingly, the policy learned by the symmetry aware control agent drives the system toward an equilibrium state of the forced KSE that is connected by continuation to an equilibrium of the unforced KSE, despite having been given no explicit information regarding its existence. That is, to achieve its goal, the RL algorithm discovers and stabilizes an equilibrium state of the system. Finally, we demonstrate that the symmetry-reduced control policy is robust to observation and actuation signal noise, as well as to system parameters it has not observed before.

摘要

深度强化学习(RL)是一种数据驱动的无模型方法,能够为高维系统中的宏观目标发现复杂的控制策略,这使得其在流量控制方面的应用前景广阔。许多人们感兴趣的流量控制系统都具有对称性,如果忽略这些对称性,可能会显著抑制简单深度强化学习方法的学习和性能。我们使用一个由Kuramoto-Sivashinsky方程(KSE)、等距致动器组成的测试平台,并以最小化耗散和功率成本为目标,证明了通过将深度强化学习问题转移到对称性降低的空间,我们可以缓解简单应用深度强化学习时固有的局限性。我们证明,与简单深度强化学习找到的策略相比,对称性降低的深度强化学习提高了数据效率以及控制策略的有效性。有趣的是,由对称性感知控制智能体学习到的策略将系统驱动到受迫KSE的一个平衡状态,该平衡状态通过连续与未受迫KSE 的一个平衡状态相连接,尽管没有给它关于其存在的明确信息就做到了这一点。也就是说,为了实现其目标,强化学习算法发现并稳定了系统的一个平衡状态。最后,我们证明,对称性降低的控制策略对观测和驱动信号噪声以及它之前未观测到的系统参数具有鲁棒性。

相似文献

1
Symmetry reduction for deep reinforcement learning active control of chaotic spatiotemporal dynamics.用于混沌时空动力学深度强化学习主动控制的对称性约化
Phys Rev E. 2021 Jul;104(1-1):014210. doi: 10.1103/PhysRevE.104.014210.
2
Control of chaotic systems by deep reinforcement learning.基于深度强化学习的混沌系统控制
Proc Math Phys Eng Sci. 2019 Nov;475(2231):20190351. doi: 10.1098/rspa.2019.0351. Epub 2019 Nov 6.
3
Continuous action deep reinforcement learning for propofol dosing during general anesthesia.全身麻醉期间丙泊酚给药的连续动作深度强化学习
Artif Intell Med. 2022 Jan;123:102227. doi: 10.1016/j.artmed.2021.102227. Epub 2021 Dec 2.
4
Characterizing Motor Control of Mastication With Soft Actor-Critic.基于软演员评论家算法的咀嚼运动控制特性分析
Front Hum Neurosci. 2020 May 26;14:188. doi: 10.3389/fnhum.2020.00188. eCollection 2020.
5
Restoring chaos using deep reinforcement learning.利用深度强化学习恢复混沌。
Chaos. 2020 Mar;30(3):031102. doi: 10.1063/5.0002047.
6
Off-Policy Reinforcement Learning for Synchronization in Multiagent Graphical Games.多智能体图博弈中的非策略强化学习同步。
IEEE Trans Neural Netw Learn Syst. 2017 Oct;28(10):2434-2445. doi: 10.1109/TNNLS.2016.2609500. Epub 2017 Apr 17.
7
Reinforcement learning for intensive care medicine: actionable clinical insights from novel approaches to reward shaping and off-policy model evaluation.重症医学中的强化学习:奖励塑造和离策略模型评估新方法带来的可操作临床见解。
Intensive Care Med Exp. 2024 Mar 25;12(1):32. doi: 10.1186/s40635-024-00614-x.
8
Sample Efficient Deep Reinforcement Learning With Online State Abstraction and Causal Transformer Model Prediction.基于在线状态抽象和因果变压器模型预测的样本高效深度强化学习
IEEE Trans Neural Netw Learn Syst. 2024 Nov;35(11):16574-16588. doi: 10.1109/TNNLS.2023.3296642. Epub 2024 Oct 29.
9
Data-Driven Robust Control of Discrete-Time Uncertain Linear Systems via Off-Policy Reinforcement Learning.基于离策略强化学习的离散时间不确定线性系统的数据驱动鲁棒控制
IEEE Trans Neural Netw Learn Syst. 2019 Dec;30(12):3735-3747. doi: 10.1109/TNNLS.2019.2897814. Epub 2019 Mar 1.
10
Cooperative Deep Reinforcement Learning for Large-Scale Traffic Grid Signal Control.用于大规模交通网格信号控制的协作深度强化学习
IEEE Trans Cybern. 2020 Jun;50(6):2687-2700. doi: 10.1109/TCYB.2019.2904742. Epub 2019 Mar 29.