• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

H Static Output-Feedback Control Design for Discrete-Time Systems Using Reinforcement Learning.

作者信息

Valadbeigi Amir Parviz, Sedigh Ali Khaki, Lewis F L

出版信息

IEEE Trans Neural Netw Learn Syst. 2020 Feb;31(2):396-406. doi: 10.1109/TNNLS.2019.2901889. Epub 2019 Apr 19.

DOI:10.1109/TNNLS.2019.2901889
PMID:31021775
Abstract

This paper provides necessary and sufficient conditions for the existence of the static output-feedback (OPFB) solution to the H control problem for linear discrete-time systems. It is shown that the solution of the static OPFB H control is a Nash equilibrium point. Furthermore, a Q-learning algorithm is developed to find the H OPFB solution online using data measured along the system trajectories and without knowing the system matrices. This is achieved by solving a game algebraic Riccati equation online and using the measured data. A simulation example shows the effectiveness of the proposed method.

摘要

相似文献

1
H Static Output-Feedback Control Design for Discrete-Time Systems Using Reinforcement Learning.
IEEE Trans Neural Netw Learn Syst. 2020 Feb;31(2):396-406. doi: 10.1109/TNNLS.2019.2901889. Epub 2019 Apr 19.
2
Optimal Output-Feedback Control of Unknown Continuous-Time Linear Systems Using Off-policy Reinforcement Learning.基于无策略强化学习的未知连续时间线性系统最优输出反馈控制。
IEEE Trans Cybern. 2016 Nov;46(11):2401-2410. doi: 10.1109/TCYB.2015.2477810. Epub 2016 Sep 22.
3
Data-Driven H Optimal Output Feedback Control for Linear Discrete-Time Systems Based on Off-Policy Q-Learning.基于离策略Q学习的线性离散时间系统数据驱动H最优输出反馈控制
IEEE Trans Neural Netw Learn Syst. 2023 Jul;34(7):3553-3567. doi: 10.1109/TNNLS.2021.3112457. Epub 2023 Jul 6.
4
Fuzzy H Control of Discrete-Time Nonlinear Markov Jump Systems via a Novel Hybrid Reinforcement Q-Learning Method.基于一种新型混合强化Q学习方法的离散时间非线性马尔可夫跳跃系统的模糊H控制
IEEE Trans Cybern. 2023 Nov;53(11):7380-7391. doi: 10.1109/TCYB.2022.3220537. Epub 2023 Oct 17.
5
Inverse Q-Learning Using Input-Output Data.使用输入输出数据的逆Q学习
IEEE Trans Cybern. 2024 Feb;54(2):728-738. doi: 10.1109/TCYB.2023.3338197. Epub 2024 Jan 17.
6
Online adaptive policy learning algorithm for H∞ state feedback control of unknown affine nonlinear discrete-time systems.用于未知仿射非线性离散时间系统 H∞状态反馈控制的在线自适应策略学习算法。
IEEE Trans Cybern. 2014 Dec;44(12):2706-18. doi: 10.1109/TCYB.2014.2313915. Epub 2014 Jul 28.
7
Data-Driven H∞ Output Consensus for Heterogeneous Multiagent Systems Under Switching Topology via Reinforcement Learning.基于强化学习的切换拓扑下异构多智能体系统的数据驱动H∞输出一致性
IEEE Trans Cybern. 2024 Dec;54(12):7865-7876. doi: 10.1109/TCYB.2024.3419056. Epub 2024 Nov 27.
8
Output Feedback Q-Learning Control for the Discrete-Time Linear Quadratic Regulator Problem.离散时间线性二次调节器问题的输出反馈Q学习控制
IEEE Trans Neural Netw Learn Syst. 2019 May;30(5):1523-1536. doi: 10.1109/TNNLS.2018.2870075. Epub 2018 Oct 8.
9
Dynamic output feedback H control for 2D fuzzy systems.二维模糊系统的动态输出反馈 H 控制。
ISA Trans. 2019 Dec;95:1-10. doi: 10.1016/j.isatra.2018.12.029. Epub 2018 Dec 21.
10
An iterative Q-learning based global consensus of discrete-time saturated multi-agent systems.基于迭代 Q 学习的离散饱和多智能体系统全局一致性。
Chaos. 2019 Oct;29(10):103127. doi: 10.1063/1.5120106.