Suppr超能文献

正态分布的回归:基于变分自编码器的灵活深度持续学习

Return of the normal distribution: Flexible deep continual learning with variational auto-encoders.

作者信息

Hong Yongwon, Mundt Martin, Park Sungho, Uh Yungjung, Byun Hyeran

机构信息

Department of Computer Science, Yonsei University, Seoul, Republic of Korea.

Department of Computer Science, TU Darmstadt and Hessian Center for Artificial Intelligence (hessian.AI), Darmstadt, Germany.

出版信息

Neural Netw. 2022 Oct;154:397-412. doi: 10.1016/j.neunet.2022.07.016. Epub 2022 Jul 21.

Abstract

Learning continually from sequentially arriving data has been a long standing challenge in machine learning. An emergent body of deep learning literature suggests various solutions, through introduction of significant simplifications to the problem statement. As a consequence of a growing focus on particular tasks and their respective benchmark assumptions, these efforts are thus becoming increasingly tailored to specific settings. Whereas approaches that leverage Variational Bayesian techniques seem to provide a more general perspective of key continual learning mechanisms, they however entail their own caveats. Inspired by prior theoretical work on solving the prevalent mismatch between prior and aggregate posterior in deep generative models, we return to a generic variational auto-encoder based formulation and investigate its utility for continual learning. Specifically, we propose to adapt a two-stage training framework towards a context conditioned variant for continual learning, where we then formulate mechanisms to alleviate catastrophic forgetting through choices of generative rehearsal or well-motivated extraction of data exemplar subsets. Although the proposed generic two-stage variational auto-encoder is not tailored towards a particular task and allows for flexible amounts of supervision, we empirically demonstrate it to surpass task-tailored methods in both supervised classification, as well as unsupervised representation learning.

摘要

在机器学习中,持续从顺序到达的数据中学习一直是一个长期存在的挑战。大量新兴的深度学习文献通过对问题陈述进行重大简化,提出了各种解决方案。由于越来越关注特定任务及其各自的基准假设,这些努力因此越来越针对特定设置进行定制。虽然利用变分贝叶斯技术的方法似乎提供了对关键持续学习机制更全面的视角,但它们也有自身的问题。受先前关于解决深度生成模型中先验与总体后验之间普遍不匹配的理论工作的启发,我们回到基于通用变分自编码器的公式,并研究其在持续学习中的效用。具体而言,我们建议将两阶段训练框架调整为用于持续学习的上下文条件变体,然后我们制定机制,通过生成性排练的选择或合理提取数据样本子集来减轻灾难性遗忘。虽然所提出的通用两阶段变分自编码器并非针对特定任务进行定制,并且允许灵活的监督量,但我们通过实验证明,它在监督分类以及无监督表示学习方面都优于针对特定任务的方法。

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验