Mostafa Hesham, Ramesh Vishwajith, Cauwenberghs Gert
Institute for Neural Computation, University of California, San Diego, San Diego, CA, United States.
Department of Bioengineering, University of California, San Diego, San Diego, CA, United States.
Front Neurosci. 2018 Aug 31;12:608. doi: 10.3389/fnins.2018.00608. eCollection 2018.
Error backpropagation is a highly effective mechanism for learning high-quality hierarchical features in deep networks. Updating the features or weights in one layer, however, requires waiting for the propagation of error signals from higher layers. Learning using delayed and non-local errors makes it hard to reconcile backpropagation with the learning mechanisms observed in biological neural networks as it requires the neurons to maintain a memory of the input long enough until the higher-layer errors arrive. In this paper, we propose an alternative learning mechanism where errors are generated locally in each layer using fixed, random auxiliary classifiers. Lower layers could thus be trained independently of higher layers and training could either proceed layer by layer, or simultaneously in all layers using local error information. We address biological plausibility concerns such as weight symmetry requirements and show that the proposed learning mechanism based on fixed, broad, and random tuning of each neuron to the classification categories outperforms the biologically-motivated feedback alignment learning technique on the CIFAR10 dataset, approaching the performance of standard backpropagation. Our approach highlights a potential biological mechanism for the supervised, or task-dependent, learning of feature hierarchies. In addition, we show that it is well suited for learning deep networks in custom hardware where it can drastically reduce memory traffic and data communication overheads. Code used to run all learning experiments is available under https://gitlab.com/hesham-mostafa/learning-using-local-erros.git.
误差反向传播是一种在深度网络中学习高质量层次特征的高效机制。然而,更新某一层的特征或权重需要等待来自更高层的误差信号传播过来。使用延迟和非局部误差进行学习使得反向传播难以与生物神经网络中观察到的学习机制相协调,因为这要求神经元将输入的记忆保持足够长的时间,直到更高层的误差到来。在本文中,我们提出了一种替代的学习机制,其中误差在每一层中使用固定的随机辅助分类器进行局部生成。因此,较低层可以独立于较高层进行训练,训练可以逐层进行,或者使用局部误差信息在所有层中同时进行。我们解决了诸如权重对称性要求等生物学合理性问题,并表明基于对每个神经元进行固定、广泛和随机调谐以适应分类类别的所提出的学习机制在CIFAR10数据集上优于具有生物学动机的反馈对齐学习技术,接近标准反向传播的性能。我们的方法突出了一种用于监督学习或任务依赖学习特征层次结构的潜在生物学机制。此外,我们表明它非常适合在定制硬件中学习深度网络,在那里它可以大幅减少内存流量和数据通信开销。用于运行所有学习实验的代码可在https://gitlab.com/hesham-mostafa/learning-using-local-erros.git获取。