Department of Neurobiology and Behavior, University of California, Irvine, CA 92697.
Canadian Centre for Behavioural Neuroscience, The University of Lethbridge, Lethbridge, Alberta T1K 3M4, Canada.
Proc Natl Acad Sci U S A. 2022 Jul 5;119(27):e2115229119. doi: 10.1073/pnas.2115229119. Epub 2022 Jun 27.
Understanding how the brain learns throughout a lifetime remains a long-standing challenge. In artificial neural networks (ANNs), incorporating novel information too rapidly results in catastrophic interference, i.e., abrupt loss of previously acquired knowledge. Complementary Learning Systems Theory (CLST) suggests that new memories can be gradually integrated into the neocortex by interleaving new memories with existing knowledge. This approach, however, has been assumed to require interleaving all existing knowledge every time something new is learned, which is implausible because it is time-consuming and requires a large amount of data. We show that deep, nonlinear ANNs can learn new information by interleaving only a subset of old items that share substantial representational similarity with the new information. By using such similarity-weighted interleaved learning (SWIL), ANNs can learn new information rapidly with a similar accuracy level and minimal interference, while using a much smaller number of old items presented per epoch (fast and data-efficient). SWIL is shown to work with various standard classification datasets (Fashion-MNIST, CIFAR10, and CIFAR100), deep neural network architectures, and in sequential learning frameworks. We show that data efficiency and speedup in learning new items are increased roughly proportionally to the number of nonoverlapping classes stored in the network, which implies an enormous possible speedup in human brains, which encode a high number of separate categories. Finally, we propose a theoretical model of how SWIL might be implemented in the brain.
理解大脑如何在一生中学习仍然是一个长期存在的挑战。在人工神经网络 (ANNs) 中,过快地引入新信息会导致灾难性干扰,即先前获得的知识突然丢失。互补学习系统理论 (CLST) 表明,新记忆可以通过将新记忆与现有知识交织在一起,逐渐整合到新皮层中。然而,这种方法被假设为每次学习新内容时都需要交织所有现有的知识,这是不可行的,因为它既耗时又需要大量的数据。我们表明,深度非线性 ANNs 可以通过仅交织与新信息具有大量表示相似性的旧项目的子集来学习新信息。通过使用这种基于相似性加权的交织学习 (SWIL),ANNs 可以以相似的准确性水平和最小的干扰快速学习新信息,同时每个时期使用的旧项目数量更少(快速且高效)。SWIL 被证明可用于各种标准分类数据集(Fashion-MNIST、CIFAR10 和 CIFAR100)、深度神经网络架构以及顺序学习框架。我们表明,学习新项的数据效率和加速与网络中存储的非重叠类别的数量大致成正比,这意味着人类大脑可能会有巨大的加速潜力,因为人类大脑编码了大量独立的类别。最后,我们提出了一个关于 SWIL 如何在大脑中实现的理论模型。