Wang Liyuan, Lei Bo, Li Qian, Su Hang, Zhu Jun, Zhong Yi
IEEE Trans Neural Netw Learn Syst. 2022 May;33(5):1925-1934. doi: 10.1109/TNNLS.2021.3111019. Epub 2022 May 2.
Continual acquisition of novel experience without interfering with previously learned knowledge, i.e., continual learning, is critical for artificial neural networks, while limited by catastrophic forgetting. A neural network adjusts its parameters when learning a new task but then fails to conduct the old tasks well. By contrast, the biological brain can effectively address catastrophic forgetting through consolidating memories as more specific or more generalized forms to complement each other, which is achieved in the interplay of the hippocampus and neocortex, mediated by the prefrontal cortex. Inspired by such a brain strategy, we propose a novel approach named triple-memory networks (TMNs) for continual learning. TMNs model the interplay of the three brain regions as a triple-network architecture of generative adversarial networks (GANs). The input information is encoded as specific representations of data distributions in a generator, or generalized knowledge of solving tasks in a discriminator and a classifier, with implementing appropriate brain-inspired algorithms to alleviate catastrophic forgetting in each module. Particularly, the generator replays generated data of the learned tasks to the discriminator and the classifier, both of which are implemented with a weight consolidation regularizer to complement the lost information in the generation process. TMNs achieve the state-of-the-art performance of generative memory replay on a variety of class-incremental learning benchmarks on MNIST, SVHN, CIFAR-10, and ImageNet-50.
在不干扰先前所学知识的情况下持续获取新经验,即持续学习,对人工神经网络至关重要,但却受到灾难性遗忘的限制。神经网络在学习新任务时会调整其参数,但随后却无法很好地执行旧任务。相比之下,生物大脑可以通过将记忆巩固为更具体或更通用的形式来相互补充,从而有效地解决灾难性遗忘问题,这是在海马体和新皮层的相互作用中实现的,由前额叶皮层介导。受这种大脑策略的启发,我们提出了一种名为三重记忆网络(TMNs)的新颖方法用于持续学习。TMNs将三个脑区的相互作用建模为生成对抗网络(GANs)的三重网络架构。输入信息在生成器中被编码为数据分布的特定表示,或在判别器和分类器中被编码为解决任务的通用知识,并通过实施适当的受大脑启发的算法来减轻每个模块中的灾难性遗忘。特别地,生成器将所学任务的生成数据重放给判别器和分类器,这两者都通过权重巩固正则化器来实现,以补充生成过程中丢失的信息。TMNs在MNIST、SVHN、CIFAR - 10和ImageNet - 50等各种类增量学习基准上实现了生成性记忆重放的最优性能。