Yu Yao-Chi, Narayanan Vignesh, Ching ShiNung, Li Jr-Shin
Department of Electrical and Systems Engineering, Washington University in St. Louis, St. Louis, MO, 63130, USA.
Division of Biology and Biomedical Sciences, Washington University in St. Louis, St. Louis, MO, 63130, USA.
Proc Am Control Conf. 2020 Jul;2020:4028-4033. doi: 10.23919/acc45564.2020.9147426. Epub 2020 Jul 27.
Controlling a population of neurons with one or a few control signals is challenging due to the severely underactuated nature of the control system and the inherent nonlinear dynamics of the neurons that are typically unknown. Control strategies that incorporate deep neural networks and machine learning techniques directly use data to learn a sequence of control actions for targeted manipulation of a population of neurons. However, these learning strategies inherently assume that perfect feedback data from each neuron at every sampling instant are available, and do not scale gracefully as the number of neurons in the population increases. As a result, the learning models need to be retrained whenever such a change occurs. In this work, we propose a learning strategy to design a control sequence by using population-level aggregated measurements and incorporate reinforcement learning techniques to find a (bounded, piecewise constant) control policy that fulfills the given control task. We demonstrate the feasibility of the proposed approach using numerical experiments on a finite population of nonlinear dynamical systems and canonical phase models that are widely used in neuroscience.
由于控制系统严重欠驱动的特性以及通常未知的神经元固有非线性动力学,用一个或几个控制信号来控制一群神经元具有挑战性。结合深度神经网络和机器学习技术的控制策略直接利用数据来学习一系列控制动作,以对一群神经元进行有针对性的操纵。然而,这些学习策略本质上假设在每个采样时刻都能获得来自每个神经元的完美反馈数据,并且随着群体中神经元数量的增加,其扩展性不佳。因此,每当发生这种变化时,学习模型都需要重新训练。在这项工作中,我们提出一种学习策略,通过使用群体水平的聚合测量来设计控制序列,并结合强化学习技术来找到一个满足给定控制任务的(有界、分段常数)控制策略。我们通过对有限数量的非线性动力系统和神经科学中广泛使用的典型相位模型进行数值实验,证明了所提出方法的可行性。