Ma Hailan, Dong Daoyi, Ding Steven X, Chen Chunlin
IEEE Trans Neural Netw Learn Syst. 2023 Nov;34(11):8852-8865. doi: 10.1109/TNNLS.2022.3153502. Epub 2023 Oct 27.
Deep reinforcement learning (DRL) has been recognized as an efficient technique to design optimal strategies for different complex systems without prior knowledge of the control landscape. To achieve a fast and precise control for quantum systems, we propose a novel DRL approach by constructing a curriculum consisting of a set of intermediate tasks defined by fidelity thresholds, where the tasks among a curriculum can be statically determined before the learning process or dynamically generated during the learning process. By transferring knowledge between two successive tasks and sequencing tasks according to their difficulties, the proposed curriculum-based DRL (CDRL) method enables the agent to focus on easy tasks in the early stage, then move onto difficult tasks, and eventually approaches the final task. Numerical comparison with the traditional methods [gradient method (GD), genetic algorithm (GA), and several other DRL methods] demonstrates that CDRL exhibits improved control performance for quantum systems and also provides an efficient way to identify optimal strategies with few control pulses.
深度强化学习(DRL)已被公认为是一种有效的技术,可在无需事先了解控制格局的情况下为不同复杂系统设计最优策略。为了实现对量子系统的快速精确控制,我们提出了一种新颖的深度强化学习方法,通过构建一个由一组由保真度阈值定义的中间任务组成的课程,其中课程中的任务可以在学习过程之前静态确定,也可以在学习过程中动态生成。通过在两个连续任务之间传递知识并根据任务难度对任务进行排序,所提出的基于课程的深度强化学习(CDRL)方法使智能体能够在早期专注于简单任务,然后转向困难任务,并最终接近最终任务。与传统方法[梯度法(GD)、遗传算法(GA)和其他几种深度强化学习方法]的数值比较表明,CDRL对量子系统具有改进的控制性能,并且还提供了一种以较少控制脉冲识别最优策略的有效方法。