• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

相似文献

1
Data-Driven Self-Triggered Control for Networked Motor Control Systems Using RNNs and Pre-Training: A Hierarchical Reinforcement Learning Framework.使用循环神经网络和预训练的网络电机控制系统的数据驱动自触发控制:一种分层强化学习框架
Sensors (Basel). 2024 Mar 20;24(6):1986. doi: 10.3390/s24061986.
2
A priority experience replay actor-critic algorithm using self-attention mechanism for strategy optimization of discrete problems.一种使用自注意力机制的优先经验回放演员-评论家算法,用于离散问题的策略优化。
PeerJ Comput Sci. 2024 Jun 28;10:e2161. doi: 10.7717/peerj-cs.2161. eCollection 2024.
3
Meta attention for Off-Policy Actor-Critic.用于离策略演员-评论家的元注意力机制
Neural Netw. 2023 Jun;163:86-96. doi: 10.1016/j.neunet.2023.03.024. Epub 2023 Mar 28.
4
Deep Reinforcement Learning on Autonomous Driving Policy With Auxiliary Critic Network.基于辅助评论家网络的自动驾驶策略深度强化学习
IEEE Trans Neural Netw Learn Syst. 2023 Jul;34(7):3680-3690. doi: 10.1109/TNNLS.2021.3116063. Epub 2023 Jul 6.
5
Asynchronous learning for actor-critic neural networks and synchronous triggering for multiplayer system.异步学习的演员-批评神经网络和同步触发的多人系统。
ISA Trans. 2022 Oct;129(Pt B):295-308. doi: 10.1016/j.isatra.2022.02.007. Epub 2022 Feb 10.
6
IHG-MA: Inductive heterogeneous graph multi-agent reinforcement learning for multi-intersection traffic signal control.IHG-MA:用于多交叉口交通信号控制的归纳异质图多智能体强化学习。
Neural Netw. 2021 Jul;139:265-277. doi: 10.1016/j.neunet.2021.03.015. Epub 2021 Mar 22.
7
Target Tracking Control of a Biomimetic Underwater Vehicle Through Deep Reinforcement Learning.通过深度强化学习的仿生水下航行器目标跟踪控制。
IEEE Trans Neural Netw Learn Syst. 2022 Aug;33(8):3741-3752. doi: 10.1109/TNNLS.2021.3054402. Epub 2022 Aug 3.
8
Deep Deterministic Policy Gradient With Compatible Critic Network.具有兼容批评网络的深度确定性策略梯度
IEEE Trans Neural Netw Learn Syst. 2023 Aug;34(8):4332-4344. doi: 10.1109/TNNLS.2021.3117790. Epub 2023 Aug 4.
9
An actor-critic framework based on deep reinforcement learning for addressing flexible job shop scheduling problems.一种基于深度强化学习的演员-评论家框架,用于解决柔性作业车间调度问题。
Math Biosci Eng. 2024 Jan;21(1):1445-1471. doi: 10.3934/mbe.2024062. Epub 2022 Dec 28.
10
Optimal Policy of Multiplayer Poker via Actor-Critic Reinforcement Learning.通过演员-评论家强化学习实现多人扑克的最优策略
Entropy (Basel). 2022 May 30;24(6):774. doi: 10.3390/e24060774.

本文引用的文献

1
Learning Self-Triggered Controllers With Gaussian Processes.使用高斯过程学习自触发控制器。
IEEE Trans Cybern. 2021 Dec;51(12):6294-6304. doi: 10.1109/TCYB.2020.2980048. Epub 2021 Dec 22.
2
Policy Gradient Adaptive Dynamic Programming for Data-Based Optimal Control.基于数据的最优控制的策略梯度自适应动态规划。
IEEE Trans Cybern. 2017 Oct;47(10):3341-3354. doi: 10.1109/TCYB.2016.2623859. Epub 2016 Nov 22.

使用循环神经网络和预训练的网络电机控制系统的数据驱动自触发控制:一种分层强化学习框架

Data-Driven Self-Triggered Control for Networked Motor Control Systems Using RNNs and Pre-Training: A Hierarchical Reinforcement Learning Framework.

作者信息

Chen Wei, Wan Haiying, Luan Xiaoli, Liu Fei

机构信息

Key Laboratory of Advanced Process Control for Light Industry, Institute of Automation, Jiangnan University, Wuxi 214122, China.

出版信息

Sensors (Basel). 2024 Mar 20;24(6):1986. doi: 10.3390/s24061986.

DOI:10.3390/s24061986
PMID:38544250
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10974179/
Abstract

This paper introduces a novel data-driven self-triggered control approach based on a hierarchical reinforcement learning framework in networked motor control systems. This approach divides the self-triggered control policy into higher and lower layers, with the higher-level policy guiding the lower-level policy in decision-making, thereby reducing the exploration space of the lower-level policy and improving the efficiency of the learning process. The data-driven framework integrates with the dual-actor critic algorithm, using two interconnected neural networks to approximate the hierarchical policies. In this framework, we use recurrent neural networks as the network architecture for the critic, utilizing the temporal dynamics of recurrent neural networks to better capture the dependencies between costs, thus enhancing the critic network's efficiency and accuracy in approximating the multi-time cumulative cost function. Additionally, we have developed a pre-training method for the control policy networks to further improve learning efficiency. The effectiveness of our proposed method is validated through a series of numerical simulations.

摘要

本文介绍了一种基于分层强化学习框架的新型数据驱动自触发控制方法,应用于网络化电机控制系统。该方法将自触发控制策略分为高层和低层,高层策略在决策中指导低层策略,从而减少低层策略的探索空间,提高学习过程的效率。数据驱动框架与双智能体评判算法相结合,使用两个相互连接的神经网络来近似分层策略。在此框架中,我们使用循环神经网络作为评判器的网络架构,利用循环神经网络的时间动态特性更好地捕捉成本之间的依赖关系,从而提高评判器网络在近似多时刻累积成本函数时的效率和准确性。此外,我们还开发了一种控制策略网络的预训练方法,以进一步提高学习效率。通过一系列数值模拟验证了我们所提方法的有效性。