Luo Yong, Yin Liancheng, Bai Wenchao, Mao Keming
College of Software, Northeastern University, Shenyang 110004, China.
Entropy (Basel). 2020 Oct 22;22(11):1190. doi: 10.3390/e22111190.
As a special case of machine learning, incremental learning can acquire useful knowledge from incoming data continuously while it does not need to access the original data. It is expected to have the ability of memorization and it is regarded as one of the ultimate goals of artificial intelligence technology. However, incremental learning remains a long term challenge. Modern deep neural network models achieve outstanding performance on stationary data distributions with batch training. This restriction leads to catastrophic forgetting for incremental learning scenarios since the distribution of incoming data is unknown and has a highly different probability from the old data. Therefore, a model must be both plastic to acquire new knowledge and stable to consolidate existing knowledge. This review aims to draw a systematic review of the state of the art of incremental learning methods. Published reports are selected from Web of Science, IEEEXplore, and DBLP databases up to May 2020. Each paper is reviewed according to the types: architectural strategy, regularization strategy and rehearsal and pseudo-rehearsal strategy. We compare and discuss different methods. Moreover, the development trend and research focus are given. It is concluded that incremental learning is still a hot research area and will be for a long period. More attention should be paid to the exploration of both biological systems and computational models.
作为机器学习的一种特殊情况,增量学习可以在无需访问原始数据的情况下,持续从传入数据中获取有用知识。它被期望具备记忆能力,并且被视为人工智能技术的最终目标之一。然而,增量学习仍然是一项长期挑战。现代深度神经网络模型在批量训练的固定数据分布上取得了出色的性能。这种限制导致在增量学习场景中出现灾难性遗忘,因为传入数据的分布是未知的,且与旧数据的概率有很大差异。因此,一个模型必须既具有可塑性以获取新知识,又具有稳定性以巩固现有知识。本综述旨在对增量学习方法的现状进行系统综述。从截至2020年5月的Web of Science、IEEE Xplore和DBLP数据库中选取已发表的报告。每篇论文按照架构策略、正则化策略以及排练和伪排练策略的类型进行综述。我们对不同方法进行比较和讨论。此外,还给出了发展趋势和研究重点。得出的结论是,增量学习仍然是一个热门研究领域,并且在很长一段时间内都会如此。应更加关注对生物系统和计算模型的探索。