Donge Vrushabh S, Lian Bosen, Lewis Frank L, Davoudi Ali
IEEE Trans Cybern. 2024 Mar;54(3):1391-1402. doi: 10.1109/TCYB.2023.3324601. Epub 2024 Feb 9.
This article proposes a data-efficient model-free reinforcement learning (RL) algorithm using Koopman operators for complex nonlinear systems. A high-dimensional data-driven optimal control of the nonlinear system is developed by lifting it into the linear system model. We use a data-driven model-based RL framework to derive an off-policy Bellman equation. Building upon this equation, we deduce the data-efficient RL algorithm, which does not need a Koopman-built linear system model. This algorithm preserves dynamic information while reducing the required data for optimal control learning. Numerical and theoretical analyses of the Koopman eigenfunctions for dataset truncation are discussed in the proposed model-free data-efficient RL algorithm. We validate our framework on the excitation control of the power system.
本文提出了一种使用库普曼算子的、适用于复杂非线性系统的数据高效免模型强化学习(RL)算法。通过将非线性系统提升到线性系统模型,开发了一种高维数据驱动的非线性系统最优控制方法。我们使用基于数据驱动模型的RL框架来推导离策略贝尔曼方程。基于此方程,我们推导出了数据高效RL算法,该算法不需要库普曼构建的线性系统模型。该算法在减少最优控制学习所需数据的同时保留了动态信息。在所提出的免模型数据高效RL算法中讨论了用于数据集截断的库普曼本征函数的数值和理论分析。我们在电力系统的励磁控制上验证了我们的框架。