Song Yuhang, Lukasiewicz Thomas, Xu Zhenghua, Bogacz Rafal
Department of Computer Science, University of Oxford, UK.
State Key Laboratory of Reliability and Intelligence of Electrical Equipment, Hebei University of Technology, Tianjin, China.
Adv Neural Inf Process Syst. 2020;33:22566-22579.
Backpropagation (BP) has been the most successful algorithm used to train artificial neural networks. However, there are several gaps between BP and learning in biologically plausible neuronal networks of the brain (learning in the brain, or simply BL, for short), in particular, (1) it has been unclear to date, if BP can be implemented exactly via BL, (2) there is a lack of local plasticity in BP, i.e., weight updates require information that is not locally available, while BL utilizes only locally available information, and (3) there is a lack of autonomy in BP, i.e., some external control over the neural network is required (e.g., switching between prediction and learning stages requires changes to dynamics and synaptic plasticity rules), while BL works fully autonomously. Bridging such gaps, i.e., understanding how BP can be approximated by BL, has been of major interest in both neuroscience and machine learning. Despite tremendous efforts, however, no previous model has bridged the gaps at a degree of demonstrating an equivalence to BP, instead, only approximations to BP have been shown. Here, we present for the first time a framework within BL that bridges the above crucial gaps. We propose a BL model that (1) produces updates of the neural weights as BP, while (2) employing local plasticity, i.e., all neurons perform only local computations, done simultaneously. We then modify it to an alternative BL model that (3) also works fully autonomously. Overall, our work provides important evidence for the debate on the long-disputed question whether the brain can perform BP.
反向传播(BP)一直是用于训练人工神经网络的最成功算法。然而,BP与大脑中具有生物学合理性的神经元网络中的学习(简称为大脑学习,或简称为BL)之间存在一些差距,特别是:(1)迄今为止尚不清楚BP是否可以通过BL精确实现;(2)BP缺乏局部可塑性,即权重更新需要局部不可用的信息,而BL仅利用局部可用信息;(3)BP缺乏自主性,即需要对神经网络进行一些外部控制(例如,在预测和学习阶段之间切换需要改变动力学和突触可塑性规则),而BL完全自主运行。弥合这些差距,即理解BL如何近似BP,一直是神经科学和机器学习领域的主要研究兴趣。然而,尽管付出了巨大努力,但以前没有模型能够在与BP等效的程度上弥合这些差距,相反,只展示了对BP的近似。在这里,我们首次提出了一个在BL框架内弥合上述关键差距的框架。我们提出了一个BL模型,该模型(1)像BP一样产生神经权重的更新,同时(2)采用局部可塑性,即所有神经元仅执行局部计算,并且是同时进行的。然后,我们将其修改为另一个BL模型,该模型(3)也能完全自主运行。总的来说,我们的工作为关于长期存在争议的问题——大脑是否能够执行BP——的辩论提供了重要证据。