• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

CKDF:用于稳健增量学习的级联知识蒸馏框架

CKDF: Cascaded Knowledge Distillation Framework for Robust Incremental Learning.

作者信息

Li Kunchi, Wan Jun, Yu Shan

出版信息

IEEE Trans Image Process. 2022;31:3825-3837. doi: 10.1109/TIP.2022.3176130. Epub 2022 Jun 2.

DOI:10.1109/TIP.2022.3176130
PMID:35609094
Abstract

Recently, owing to the superior performances, knowledge distillation-based (kd-based) methods with the exemplar rehearsal have been widely applied in class incremental learning (CIL). However, we discover that they suffer from the feature uncalibration problem, which is caused by directly transferring knowledge from the old model immediately to the new model when learning a new task. As the old model confuses the feature representations between the learned and new classes, the kd loss and the classification loss used in kd-based methods are heterogeneous. This is detrimental if we learn the existing knowledge from the old model directly in the way as in typical kd-based methods. To tackle this problem, the feature calibration network (FCN) is proposed, which is used to calibrate the existing knowledge to alleviate the feature representation confusion of the old model. In addition, to relieve the task-recency bias of FCN caused by the limited storage memory in CIL, we propose a novel image-feature hybrid sample rehearsal strategy to train FCN by splitting the memory budget to store the image-and-feature exemplars of the previous tasks. As feature embeddings of images have much lower-dimensions, this allows us to store more samples to train FCN. Based on these two improvements, we propose the Cascaded Knowledge Distillation Framework (CKDF) including three main stages. The first stage is used to train FCN to calibrate the existing knowledge of the old model. Then, the new model is trained simultaneously by transferring knowledge from the calibrated teacher model through the knowledge distillation strategy and learning new classes. Finally, after completing the new task learning, the feature exemplars of previous tasks are updated. Importantly, we demonstrate that the proposed CKDF is a general framework that can be applied to various kd-based methods. Experimental results show that our method achieves state-of-the-art performances on several CIL benchmarks.

摘要

最近,由于性能优越,基于知识蒸馏(KD)并结合样本排练的方法已广泛应用于类别增量学习(CIL)。然而,我们发现它们存在特征未校准问题,这是在学习新任务时直接将旧模型的知识立即转移到新模型所导致的。由于旧模型混淆了已学类别和新类别之间的特征表示,基于KD的方法中使用的KD损失和分类损失是异构的。如果我们像在典型的基于KD的方法中那样直接从旧模型学习现有知识,这是有害的。为了解决这个问题,我们提出了特征校准网络(FCN),用于校准现有知识,以减轻旧模型的特征表示混淆。此外,为了缓解CIL中有限存储内存导致的FCN任务近期偏差,我们提出了一种新颖的图像-特征混合样本排练策略,通过分割内存预算来存储先前任务的图像和特征样本,从而训练FCN。由于图像的特征嵌入维度低得多,这使我们能够存储更多样本以训练FCN。基于这两项改进,我们提出了级联知识蒸馏框架(CKDF),包括三个主要阶段。第一阶段用于训练FCN以校准旧模型的现有知识。然后,通过知识蒸馏策略从校准后的教师模型转移知识并学习新类别,同时训练新模型。最后,在完成新任务学习后,更新先前任务的特征样本。重要的是,我们证明了所提出的CKDF是一个通用框架,可应用于各种基于KD的方法。实验结果表明,我们的方法在几个CIL基准测试中取得了领先的性能。

相似文献

1
CKDF: Cascaded Knowledge Distillation Framework for Robust Incremental Learning.CKDF:用于稳健增量学习的级联知识蒸馏框架
IEEE Trans Image Process. 2022;31:3825-3837. doi: 10.1109/TIP.2022.3176130. Epub 2022 Jun 2.
2
Dual Balanced Class-Incremental Learning With im-Softmax and Angular Rectification.基于im-Softmax和角度校正的双平衡类增量学习
IEEE Trans Neural Netw Learn Syst. 2025 Mar;36(3):4437-4447. doi: 10.1109/TNNLS.2024.3368341. Epub 2025 Feb 28.
3
Multi-granularity knowledge distillation and prototype consistency regularization for class-incremental learning.多粒度知识蒸馏和原型一致性正则化的类增量学习。
Neural Netw. 2023 Jul;164:617-630. doi: 10.1016/j.neunet.2023.05.006. Epub 2023 May 11.
4
Memory-Efficient Class-Incremental Learning for Image Classification.用于图像分类的内存高效类增量学习
IEEE Trans Neural Netw Learn Syst. 2022 Oct;33(10):5966-5977. doi: 10.1109/TNNLS.2021.3072041. Epub 2022 Oct 5.
5
Memory-Replay Knowledge Distillation.记忆重放知识蒸馏。
Sensors (Basel). 2021 Apr 15;21(8):2792. doi: 10.3390/s21082792.
6
A class-incremental learning approach for learning feature-compatible embeddings.一种用于学习特征兼容嵌入的类增量学习方法。
Neural Netw. 2024 Dec;180:106685. doi: 10.1016/j.neunet.2024.106685. Epub 2024 Aug 31.
7
Light-M: An efficient lightweight medical image segmentation framework for resource-constrained IoMT.轻量级医学图像分割框架 Light-M,适用于资源受限的 IoMT。
Comput Biol Med. 2024 Mar;170:108088. doi: 10.1016/j.compbiomed.2024.108088. Epub 2024 Feb 3.
8
Less confidence, less forgetting: Learning with a humbler teacher in exemplar-free Class-Incremental learning.缺乏自信,遗忘更少:在无范例的 Class-Incremental 学习中,向更谦逊的教师学习。
Neural Netw. 2024 Nov;179:106513. doi: 10.1016/j.neunet.2024.106513. Epub 2024 Jul 6.
9
Class-Incremental Learning Method With Fast Update and High Retainability Based on Broad Learning System.基于广义学习系统的具有快速更新和高保持性的类增量学习方法
IEEE Trans Neural Netw Learn Syst. 2024 Aug;35(8):11332-11345. doi: 10.1109/TNNLS.2023.3259016. Epub 2024 Aug 5.
10
Uncertainty-Aware Contrastive Distillation for Incremental Semantic Segmentation.用于增量语义分割的不确定性感知对比蒸馏
IEEE Trans Pattern Anal Mach Intell. 2023 Feb;45(2):2567-2581. doi: 10.1109/TPAMI.2022.3163806. Epub 2023 Jan 6.