• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

通过探索海森矩阵的特征值克服持续学习中的灾难性遗忘

Overcoming Catastrophic Forgetting in Continual Learning by Exploring Eigenvalues of Hessian Matrix.

作者信息

Kong Yajing, Liu Liu, Chen Huanhuan, Kacprzyk Janusz, Tao Dacheng

出版信息

IEEE Trans Neural Netw Learn Syst. 2024 Nov;35(11):16196-16210. doi: 10.1109/TNNLS.2023.3292359. Epub 2024 Oct 29.

DOI:10.1109/TNNLS.2023.3292359
PMID:37478043
Abstract

Neural networks tend to suffer performance deterioration on previous tasks when they are applied to multiple tasks sequentially without access to previous data. The problem is commonly known as catastrophic forgetting, a significant challenge in continual learning (CL). To overcome the catastrophic forgetting, regularization-based CL methods construct a regularization-based term, which can be considered as the approximation loss function of previous tasks, to penalize the update of parameters. However, the rigorous theoretical analysis of regularization-based methods is limited. Therefore, we theoretically analyze the forgetting and the convergence properties of regularization-based methods. The theoretical results demonstrate that the upper bound of the forgetting has a relationship with the maximum eigenvalue of the Hessian matrix. Hence, to decrease the upper bound of the forgetting, we propose eiGenvalues ExplorAtion Regularization-based (GEAR) method, which explores the geometric properties of the approximation loss of prior tasks regarding the maximum eigenvalue. Extensive experimental results demonstrate that our method mitigates catastrophic forgetting and outperforms existing regularization-based methods.

摘要

当神经网络在不访问先前数据的情况下依次应用于多个任务时,它们往往会在先前任务上出现性能下降。这个问题通常被称为灾难性遗忘,是持续学习(CL)中的一个重大挑战。为了克服灾难性遗忘,基于正则化的持续学习方法构建了一个基于正则化的项,它可以被视为先前任务的近似损失函数,以惩罚参数的更新。然而,对基于正则化方法的严格理论分析是有限的。因此,我们从理论上分析了基于正则化方法的遗忘和收敛特性。理论结果表明,遗忘的上限与海森矩阵的最大特征值有关。因此,为了降低遗忘的上限,我们提出了基于特征值探索正则化(GEAR)的方法,该方法探索了先前任务近似损失关于最大特征值的几何性质。大量实验结果表明,我们的方法减轻了灾难性遗忘,并且优于现有的基于正则化的方法。

相似文献

1
Overcoming Catastrophic Forgetting in Continual Learning by Exploring Eigenvalues of Hessian Matrix.通过探索海森矩阵的特征值克服持续学习中的灾难性遗忘
IEEE Trans Neural Netw Learn Syst. 2024 Nov;35(11):16196-16210. doi: 10.1109/TNNLS.2023.3292359. Epub 2024 Oct 29.
2
Encoding primitives generation policy learning for robotic arm to overcome catastrophic forgetting in sequential multi-tasks learning.生成策略学习用于机械臂的编码基元,以克服顺序多任务学习中的灾难性遗忘。
Neural Netw. 2020 Sep;129:163-173. doi: 10.1016/j.neunet.2020.06.003. Epub 2020 Jun 5.
3
Adversarial Feature Alignment: Avoid Catastrophic Forgetting in Incremental Task Lifelong Learning.对抗性特征对齐:避免增量任务终身学习中的灾难性遗忘
Neural Comput. 2019 Nov;31(11):2266-2291. doi: 10.1162/neco_a_01232. Epub 2019 Sep 16.
4
Incremental Embedding Learning With Disentangled Representation Translation.基于解缠表示转换的增量嵌入学习
IEEE Trans Neural Netw Learn Syst. 2024 Mar;35(3):3821-3833. doi: 10.1109/TNNLS.2022.3199816. Epub 2024 Feb 29.
5
Continual Learning for Activity Recognition.持续学习的活动识别。
Annu Int Conf IEEE Eng Med Biol Soc. 2022 Jul;2022:2416-2420. doi: 10.1109/EMBC48229.2022.9871690.
6
Memory Recall: A Simple Neural Network Training Framework Against Catastrophic Forgetting.记忆召回:一种针对灾难性遗忘的简单神经网络训练框架。
IEEE Trans Neural Netw Learn Syst. 2022 May;33(5):2010-2022. doi: 10.1109/TNNLS.2021.3099700. Epub 2022 May 2.
7
Continual Learning With Knowledge Distillation: A Survey.基于知识蒸馏的持续学习:一项综述。
IEEE Trans Neural Netw Learn Syst. 2024 Oct 18;PP. doi: 10.1109/TNNLS.2024.3476068.
8
Overcoming Long-Term Catastrophic Forgetting Through Adversarial Neural Pruning and Synaptic Consolidation.通过对抗性神经修剪和突触巩固来克服长期灾难性遗忘。
IEEE Trans Neural Netw Learn Syst. 2022 Sep;33(9):4243-4256. doi: 10.1109/TNNLS.2021.3056201. Epub 2022 Aug 31.
9
Continual Learning Using Bayesian Neural Networks.贝叶斯神经网络的持续学习。
IEEE Trans Neural Netw Learn Syst. 2021 Sep;32(9):4243-4252. doi: 10.1109/TNNLS.2020.3017292. Epub 2021 Aug 31.
10
Continual learning with attentive recurrent neural networks for temporal data classification.用于时态数据分类的基于注意力循环神经网络的持续学习
Neural Netw. 2023 Jan;158:171-187. doi: 10.1016/j.neunet.2022.10.031. Epub 2022 Nov 11.

引用本文的文献

1
Tuned Compositional Feature Replays for Efficient Stream Learning.用于高效流学习的调谐组合特征回放
IEEE Trans Neural Netw Learn Syst. 2025 Feb;36(2):3300-3314. doi: 10.1109/TNNLS.2023.3344085. Epub 2025 Feb 6.