• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

通过多知识转移实现协作式知识蒸馏

Collaborative Knowledge Distillation via Multiknowledge Transfer.

作者信息

Gou Jianping, Sun Liyuan, Yu Baosheng, Du Lan, Ramamohanarao Kotagiri, Tao Dacheng

出版信息

IEEE Trans Neural Netw Learn Syst. 2024 May;35(5):6718-6730. doi: 10.1109/TNNLS.2022.3212733. Epub 2024 May 2.

DOI:10.1109/TNNLS.2022.3212733
PMID:36264723
Abstract

Knowledge distillation (KD), as an efficient and effective model compression technique, has received considerable attention in deep learning. The key to its success is about transferring knowledge from a large teacher network to a small student network. However, most existing KD methods consider only one type of knowledge learned from either instance features or relations via a specific distillation strategy, failing to explore the idea of transferring different types of knowledge with different distillation strategies. Moreover, the widely used offline distillation also suffers from a limited learning capacity due to the fixed large-to-small teacher-student architecture. In this article, we devise a collaborative KD via multiknowledge transfer (CKD-MKT) that prompts both self-learning and collaborative learning in a unified framework. Specifically, CKD-MKT utilizes a multiple knowledge transfer framework that assembles self and online distillation strategies to effectively: 1) fuse different kinds of knowledge, which allows multiple students to learn knowledge from both individual instances and instance relations, and 2) guide each other by learning from themselves using collaborative and self-learning. Experiments and ablation studies on six image datasets demonstrate that the proposed CKD-MKT significantly outperforms recent state-of-the-art methods for KD.

摘要

知识蒸馏(KD)作为一种高效的模型压缩技术,在深度学习中受到了广泛关注。其成功的关键在于将知识从大型教师网络转移到小型学生网络。然而,大多数现有的KD方法仅通过特定的蒸馏策略考虑从实例特征或关系中学到的一种类型的知识,未能探索使用不同蒸馏策略转移不同类型知识的想法。此外,由于固定的从大到小的师生架构,广泛使用的离线蒸馏也存在学习能力有限的问题。在本文中,我们设计了一种通过多知识转移的协作式KD(CKD-MKT),它在一个统一的框架中促进自学习和协作学习。具体而言,CKD-MKT利用一个多知识转移框架,该框架整合了自蒸馏和在线蒸馏策略,以有效地:1)融合不同类型的知识,使多个学生能够从单个实例和实例关系中学习知识;2)通过协作学习和自学习从自身学习中相互指导。在六个图像数据集上进行的实验和消融研究表明,所提出的CKD-MKT显著优于最近最先进的KD方法。

相似文献

1
Collaborative Knowledge Distillation via Multiknowledge Transfer.通过多知识转移实现协作式知识蒸馏
IEEE Trans Neural Netw Learn Syst. 2024 May;35(5):6718-6730. doi: 10.1109/TNNLS.2022.3212733. Epub 2024 May 2.
2
Leveraging different learning styles for improved knowledge distillation in biomedical imaging.利用不同的学习方式提高生物医学成像中的知识蒸馏效果。
Comput Biol Med. 2024 Jan;168:107764. doi: 10.1016/j.compbiomed.2023.107764. Epub 2023 Nov 30.
3
Teacher-student complementary sample contrastive distillation.师生互补样本对比蒸馏。
Neural Netw. 2024 Feb;170:176-189. doi: 10.1016/j.neunet.2023.11.036. Epub 2023 Nov 17.
4
DCCD: Reducing Neural Network Redundancy via Distillation.DCCD:通过蒸馏减少神经网络冗余
IEEE Trans Neural Netw Learn Syst. 2024 Jul;35(7):10006-10017. doi: 10.1109/TNNLS.2023.3238337. Epub 2024 Jul 8.
5
Light-M: An efficient lightweight medical image segmentation framework for resource-constrained IoMT.轻量级医学图像分割框架 Light-M,适用于资源受限的 IoMT。
Comput Biol Med. 2024 Mar;170:108088. doi: 10.1016/j.compbiomed.2024.108088. Epub 2024 Feb 3.
6
A General Dynamic Knowledge Distillation Method for Visual Analytics.一种用于视觉分析的通用动态知识蒸馏方法。
IEEE Trans Image Process. 2022 Oct 13;PP. doi: 10.1109/TIP.2022.3212905.
7
Knowledge Fusion Distillation: Improving Distillation with Multi-scale Attention Mechanisms.知识融合蒸馏:利用多尺度注意力机制改进蒸馏
Neural Process Lett. 2023 Jan 3:1-16. doi: 10.1007/s11063-022-11132-w.
8
Highlight Every Step: Knowledge Distillation via Collaborative Teaching.突出每个步骤:通过协作教学进行知识蒸馏。
IEEE Trans Cybern. 2022 Apr;52(4):2070-2081. doi: 10.1109/TCYB.2020.3007506. Epub 2022 Apr 5.
9
Knowledge Distillation Using Hierarchical Self-Supervision Augmented Distribution.使用分层自监督增强分布的知识蒸馏
IEEE Trans Neural Netw Learn Syst. 2024 Feb;35(2):2094-2108. doi: 10.1109/TNNLS.2022.3186807. Epub 2024 Feb 5.
10
Learning From Human Educational Wisdom: A Student-Centered Knowledge Distillation Method.从人类教育智慧中学习:一种以学生为中心的知识蒸馏方法。
IEEE Trans Pattern Anal Mach Intell. 2024 Jun;46(6):4188-4205. doi: 10.1109/TPAMI.2024.3354928. Epub 2024 May 7.

引用本文的文献

1
A Cross-Modal Mutual Knowledge Distillation Framework for Alzheimer's Disease Diagnosis: Addressing Incomplete Modalities.一种用于阿尔茨海默病诊断的跨模态互知识蒸馏框架:解决模态不完整问题。
IEEE Trans Autom Sci Eng. 2025;22:14218-14233. doi: 10.1109/tase.2025.3556290. Epub 2025 Mar 31.