• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

缺乏自信,遗忘更少:在无范例的 Class-Incremental 学习中,向更谦逊的教师学习。

Less confidence, less forgetting: Learning with a humbler teacher in exemplar-free Class-Incremental learning.

机构信息

National University of Defense Technology, Changsha 410000, China; State Key Laboratory of Complex & Critical Software Environment, Changsha 410000, China.

National University of Defense Technology, Changsha 410000, China; State Key Laboratory of Complex & Critical Software Environment, Changsha 410000, China.

出版信息

Neural Netw. 2024 Nov;179:106513. doi: 10.1016/j.neunet.2024.106513. Epub 2024 Jul 6.

DOI:10.1016/j.neunet.2024.106513
PMID:39018945
Abstract

Class-Incremental learning (CIL) is challenging due to catastrophic forgetting (CF), which escalates in exemplar-free scenarios. To mitigate CF, Knowledge Distillation (KD), which leverages old models as teacher models, has been widely employed in CIL. However, based on a case study, our investigation reveals that the teacher model exhibits over-confidence in unseen new samples. In this article, we conduct empirical experiments and provide theoretical analysis to investigate the over-confident phenomenon and the impact of KD in exemplar-free CIL, where access to old samples is unavailable. Building on our analysis, we propose a novel approach, Learning with Humbler Teacher, by systematically selecting an appropriate checkpoint model as a humbler teacher to mitigate CF. Furthermore, we explore utilizing the nuclear norm to obtain an appropriate temporal ensemble to enhance model stability. Notably, LwHT outperforms the state-of-the-art approach by a significant margin of 10.41%, 6.56%, and 4.31% in various settings while demonstrating superior model plasticity.

摘要

类增量学习(CIL)由于灾难性遗忘(CF)而具有挑战性,在无范例场景中,CF 会加剧。为了减轻 CF,知识蒸馏(KD)被广泛应用于 CIL 中,它利用旧模型作为教师模型。然而,根据一项案例研究,我们的调查揭示了教师模型对未见的新样本表现出过度自信。在本文中,我们进行了实证实验,并提供了理论分析,以研究无范例 CIL 中过度自信的现象和 KD 的影响,在这种情况下,无法访问旧样本。基于我们的分析,我们提出了一种新的方法,即使用更谦虚的教师进行学习(Learning with Humbler Teacher,LwHT),通过系统地选择适当的检查点模型作为更谦虚的教师来减轻 CF。此外,我们还探索利用核范数获得适当的时间集成,以增强模型稳定性。值得注意的是,LwHT 在各种设置下的表现均优于最先进的方法,其优势分别为 10.41%、6.56%和 4.31%,同时还表现出更高的模型可塑性。

相似文献

1
Less confidence, less forgetting: Learning with a humbler teacher in exemplar-free Class-Incremental learning.缺乏自信,遗忘更少:在无范例的 Class-Incremental 学习中,向更谦逊的教师学习。
Neural Netw. 2024 Nov;179:106513. doi: 10.1016/j.neunet.2024.106513. Epub 2024 Jul 6.
2
Multi-granularity knowledge distillation and prototype consistency regularization for class-incremental learning.多粒度知识蒸馏和原型一致性正则化的类增量学习。
Neural Netw. 2023 Jul;164:617-630. doi: 10.1016/j.neunet.2023.05.006. Epub 2023 May 11.
3
Memory-Efficient Class-Incremental Learning for Image Classification.用于图像分类的内存高效类增量学习
IEEE Trans Neural Netw Learn Syst. 2022 Oct;33(10):5966-5977. doi: 10.1109/TNNLS.2021.3072041. Epub 2022 Oct 5.
4
CKDF: Cascaded Knowledge Distillation Framework for Robust Incremental Learning.CKDF:用于稳健增量学习的级联知识蒸馏框架
IEEE Trans Image Process. 2022;31:3825-3837. doi: 10.1109/TIP.2022.3176130. Epub 2022 Jun 2.
5
Continual learning with attentive recurrent neural networks for temporal data classification.用于时态数据分类的基于注意力循环神经网络的持续学习
Neural Netw. 2023 Jan;158:171-187. doi: 10.1016/j.neunet.2022.10.031. Epub 2022 Nov 11.
6
Adversarial Feature Alignment: Avoid Catastrophic Forgetting in Incremental Task Lifelong Learning.对抗性特征对齐:避免增量任务终身学习中的灾难性遗忘
Neural Comput. 2019 Nov;31(11):2266-2291. doi: 10.1162/neco_a_01232. Epub 2019 Sep 16.
7
Class incremental learning of remote sensing images based on class similarity distillation.基于类相似性蒸馏的遥感图像类别增量学习
PeerJ Comput Sci. 2023 Sep 27;9:e1583. doi: 10.7717/peerj-cs.1583. eCollection 2023.
8
PAMK: Prototype Augmented Multi-Teacher Knowledge Transfer Network for Continual Zero-Shot Learning.PAMK:用于持续零样本学习的原型增强多教师知识转移网络。
IEEE Trans Image Process. 2024;33:3353-3368. doi: 10.1109/TIP.2024.3403053. Epub 2024 May 31.
9
CL3: Generalization of Contrastive Loss for Lifelong Learning.CL3:用于终身学习的对比损失的泛化
J Imaging. 2023 Nov 23;9(12):259. doi: 10.3390/jimaging9120259.
10
A comprehensive study of class incremental learning algorithms for visual tasks.面向视觉任务的类增量学习算法的综合研究。
Neural Netw. 2021 Mar;135:38-54. doi: 10.1016/j.neunet.2020.12.003. Epub 2020 Dec 8.