Ye Fei, Bors Adrian G
IEEE Trans Pattern Anal Mach Intell. 2023 May;45(5):5731-5748. doi: 10.1109/TPAMI.2022.3220928. Epub 2023 Apr 3.
Lifelong learning (LLL) represents the ability of an artificial intelligence system to learn successively a sequence of different databases. In this paper we introduce the Dynamic Self-Supervised Teacher-Student Network (D-TS), representing a more general LLL framework, where the Teacher is implemented as a dynamically expanding mixture model which automatically increases its capacity to deal with a growing number of tasks. We propose the Knowledge Discrepancy Score (KDS) criterion for measuring the relevance of the incoming information characterizing a new task when compared to the existing knowledge accumulated by the Teacher module from its previous training. The KDS ensures a light Teacher architecture while also enabling to reuse the learned knowledge whenever appropriate, accelerating the learning of given tasks. The Student module is implemented as a lightweight probabilistic generative model. We introduce a novel self-supervised learning procedure for the Student that allows to capture cross-domain latent representations from the entire knowledge accumulated by the Teacher as well as from novel data. We perform several experiments which show that D-TS can achieve the state of the art results in LLL while requiring fewer parameters than other methods.
终身学习(LLL)代表人工智能系统连续学习一系列不同数据库的能力。在本文中,我们介绍了动态自监督师生网络(D-TS),它代表了一个更通用的LLL框架,其中教师被实现为一个动态扩展的混合模型,该模型会自动增加其处理越来越多任务的能力。我们提出了知识差异分数(KDS)标准,用于衡量与教师模块从其先前训练中积累的现有知识相比,表征新任务的传入信息的相关性。KDS确保了轻量级的教师架构,同时还能够在适当的时候重用所学知识,加速给定任务的学习。学生模块被实现为一个轻量级概率生成模型。我们为学生引入了一种新颖的自监督学习过程,该过程允许从教师积累的全部知识以及新数据中捕获跨域潜在表示。我们进行了多项实验,结果表明D-TS在LLL方面可以取得优于现有技术的结果,同时所需参数比其他方法更少。