Hong Yongwon, Mundt Martin, Park Sungho, Uh Yungjung, Byun Hyeran
Department of Computer Science, Yonsei University, Seoul, Republic of Korea.
Department of Computer Science, TU Darmstadt and Hessian Center for Artificial Intelligence (hessian.AI), Darmstadt, Germany.
Neural Netw. 2022 Oct;154:397-412. doi: 10.1016/j.neunet.2022.07.016. Epub 2022 Jul 21.
Learning continually from sequentially arriving data has been a long standing challenge in machine learning. An emergent body of deep learning literature suggests various solutions, through introduction of significant simplifications to the problem statement. As a consequence of a growing focus on particular tasks and their respective benchmark assumptions, these efforts are thus becoming increasingly tailored to specific settings. Whereas approaches that leverage Variational Bayesian techniques seem to provide a more general perspective of key continual learning mechanisms, they however entail their own caveats. Inspired by prior theoretical work on solving the prevalent mismatch between prior and aggregate posterior in deep generative models, we return to a generic variational auto-encoder based formulation and investigate its utility for continual learning. Specifically, we propose to adapt a two-stage training framework towards a context conditioned variant for continual learning, where we then formulate mechanisms to alleviate catastrophic forgetting through choices of generative rehearsal or well-motivated extraction of data exemplar subsets. Although the proposed generic two-stage variational auto-encoder is not tailored towards a particular task and allows for flexible amounts of supervision, we empirically demonstrate it to surpass task-tailored methods in both supervised classification, as well as unsupervised representation learning.
在机器学习中,持续从顺序到达的数据中学习一直是一个长期存在的挑战。大量新兴的深度学习文献通过对问题陈述进行重大简化,提出了各种解决方案。由于越来越关注特定任务及其各自的基准假设,这些努力因此越来越针对特定设置进行定制。虽然利用变分贝叶斯技术的方法似乎提供了对关键持续学习机制更全面的视角,但它们也有自身的问题。受先前关于解决深度生成模型中先验与总体后验之间普遍不匹配的理论工作的启发,我们回到基于通用变分自编码器的公式,并研究其在持续学习中的效用。具体而言,我们建议将两阶段训练框架调整为用于持续学习的上下文条件变体,然后我们制定机制,通过生成性排练的选择或合理提取数据样本子集来减轻灾难性遗忘。虽然所提出的通用两阶段变分自编码器并非针对特定任务进行定制,并且允许灵活的监督量,但我们通过实验证明,它在监督分类以及无监督表示学习方面都优于针对特定任务的方法。