Song H Francis, Yang Guangyu R, Wang Xiao-Jing
Center for Neural Science, New York University, New York, New York, United States of America.
NYU-ECNU Institute of Brain and Cognitive Science, NYU Shanghai, Shanghai, China.
PLoS Comput Biol. 2016 Feb 29;12(2):e1004792. doi: 10.1371/journal.pcbi.1004792. eCollection 2016 Feb.
The ability to simultaneously record from large numbers of neurons in behaving animals has ushered in a new era for the study of the neural circuit mechanisms underlying cognitive functions. One promising approach to uncovering the dynamical and computational principles governing population responses is to analyze model recurrent neural networks (RNNs) that have been optimized to perform the same tasks as behaving animals. Because the optimization of network parameters specifies the desired output but not the manner in which to achieve this output, "trained" networks serve as a source of mechanistic hypotheses and a testing ground for data analyses that link neural computation to behavior. Complete access to the activity and connectivity of the circuit, and the ability to manipulate them arbitrarily, make trained networks a convenient proxy for biological circuits and a valuable platform for theoretical investigation. However, existing RNNs lack basic biological features such as the distinction between excitatory and inhibitory units (Dale's principle), which are essential if RNNs are to provide insights into the operation of biological circuits. Moreover, trained networks can achieve the same behavioral performance but differ substantially in their structure and dynamics, highlighting the need for a simple and flexible framework for the exploratory training of RNNs. Here, we describe a framework for gradient descent-based training of excitatory-inhibitory RNNs that can incorporate a variety of biological knowledge. We provide an implementation based on the machine learning library Theano, whose automatic differentiation capabilities facilitate modifications and extensions. We validate this framework by applying it to well-known experimental paradigms such as perceptual decision-making, context-dependent integration, multisensory integration, parametric working memory, and motor sequence generation. Our results demonstrate the wide range of neural activity patterns and behavior that can be modeled, and suggest a unified setting in which diverse cognitive computations and mechanisms can be studied.
在行为动物中同时记录大量神经元的能力,开创了一个研究认知功能背后神经回路机制的新时代。一种有前景的揭示群体反应动态和计算原理的方法是分析经过优化以执行与行为动物相同任务的循环神经网络(RNN)模型。由于网络参数的优化指定了期望的输出,但没有指定实现该输出的方式,“训练有素”的网络成为了机制假设的来源以及将神经计算与行为联系起来的数据分析的试验场。能够完全访问电路的活动和连接性,并能够随意操纵它们,使得训练有素的网络成为生物电路的便捷替代物以及理论研究的宝贵平台。然而,现有的循环神经网络缺乏基本的生物学特征,如兴奋性和抑制性单元之间的区别(戴尔原则),而如果循环神经网络要深入了解生物电路的运作,这些特征是必不可少的。此外,训练有素的网络可以实现相同的行为表现,但它们的结构和动态却有很大差异,这凸显了对循环神经网络进行探索性训练的简单灵活框架的需求。在这里,我们描述了一个基于梯度下降训练兴奋性 - 抑制性循环神经网络的框架,该框架可以纳入各种生物学知识。我们提供了一个基于机器学习库Theano的实现,其自动微分功能便于修改和扩展。我们通过将其应用于诸如感知决策、上下文依赖整合、多感官整合、参数化工作记忆和运动序列生成等著名实验范式来验证这个框架。我们的结果展示了可以建模的广泛神经活动模式和行为,并提出了一个统一的环境,在其中可以研究各种认知计算和机制。