• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于 Transformer 的目标检测知识融合

Knowledge Amalgamation for Object Detection With Transformers.

出版信息

IEEE Trans Image Process. 2023;32:2093-2106. doi: 10.1109/TIP.2023.3263105.

DOI:10.1109/TIP.2023.3263105
PMID:37023145
Abstract

Knowledge amalgamation (KA) is a novel deep model reusing task aiming to transfer knowledge from several well-trained teachers to a multi-talented and compact student. Currently, most of these approaches are tailored for convolutional neural networks (CNNs). However, there is a tendency that Transformers, with a completely different architecture, are starting to challenge the domination of CNNs in many computer vision tasks. Nevertheless, directly applying the previous KA methods to Transformers leads to severe performance degradation. In this work, we explore a more effective KA scheme for Transformer-based object detection models. Specifically, considering the architecture characteristics of Transformers, we propose to dissolve the KA into two aspects: sequence-level amalgamation (SA) and task-level amalgamation (TA). In particular, a hint is generated within the sequence-level amalgamation by concatenating teacher sequences instead of redundantly aggregating them to a fixed-size one as previous KA approaches. Besides, the student learns heterogeneous detection tasks through soft targets with efficiency in the task-level amalgamation. Extensive experiments on PASCAL VOC and COCO have unfolded that the sequence-level amalgamation significantly boosts the performance of students, while the previous methods impair the students. Moreover, the Transformer-based students excel in learning amalgamated knowledge, as they have mastered heterogeneous detection tasks rapidly and achieved superior or at least comparable performance to those of the teachers in their specializations.

摘要

知识融合(KA)是一种新颖的深度模型重用任务,旨在将知识从多个训练有素的教师转移到一个多才多艺且紧凑的学生中。目前,这些方法大多针对卷积神经网络(CNNs)进行了定制。然而,情况是,具有完全不同架构的 Transformer 开始在许多计算机视觉任务中挑战 CNN 的主导地位。尽管如此,直接将之前的 KA 方法应用于 Transformer 会导致性能严重下降。在这项工作中,我们探索了一种更有效的基于 Transformer 的目标检测模型的 KA 方案。具体来说,考虑到 Transformer 的架构特点,我们提出将 KA 分解为两个方面:序列级融合(SA)和任务级融合(TA)。具体来说,序列级融合通过串联教师序列而不是像之前的 KA 方法那样将它们冗余地聚合到一个固定大小的序列中来生成提示。此外,学生通过具有效率的软目标在任务级融合中学习异构检测任务。在 PASCAL VOC 和 COCO 上进行的广泛实验表明,序列级融合显著提高了学生的性能,而之前的方法则损害了学生的性能。此外,基于 Transformer 的学生在学习融合知识方面表现出色,因为他们能够快速掌握异构检测任务,并在专门领域中实现优于或至少可与教师相媲美的性能。

相似文献

1
Knowledge Amalgamation for Object Detection With Transformers.基于 Transformer 的目标检测知识融合
IEEE Trans Image Process. 2023;32:2093-2106. doi: 10.1109/TIP.2023.3263105.
2
Vision Transformer-based recognition of diabetic retinopathy grade.基于 Vision Transformer 的糖尿病视网膜病变分级识别。
Med Phys. 2021 Dec;48(12):7850-7863. doi: 10.1002/mp.15312. Epub 2021 Nov 16.
3
Unsupervised Pre-Training for Detection Transformers.用于检测变压器的无监督预训练
IEEE Trans Pattern Anal Mach Intell. 2023 Nov;45(11):12772-12782. doi: 10.1109/TPAMI.2022.3216514. Epub 2023 Oct 3.
4
Data-Efficient Training of Pure Vision Transformers for the Task of Chest X-ray Abnormality Detection Using Knowledge Distillation.基于知识蒸馏的数据高效训练纯视觉 Transformer 模型以实现胸部 X 射线异常检测任务
Annu Int Conf IEEE Eng Med Biol Soc. 2022 Jul;2022:1444-1447. doi: 10.1109/EMBC48229.2022.9871372.
5
Convolutions are competitive with transformers for protein sequence pretraining.卷积运算在蛋白质序列预训练方面与转换器竞争。
Cell Syst. 2024 Mar 20;15(3):286-294.e2. doi: 10.1016/j.cels.2024.01.008. Epub 2024 Feb 29.
6
Transformers-sklearn: a toolkit for medical language understanding with transformer-based models.Transformer-sklearn:一个基于 Transformer 的模型的医学语言理解工具包。
BMC Med Inform Decis Mak. 2021 Jul 30;21(Suppl 2):90. doi: 10.1186/s12911-021-01459-0.
7
A novel hybrid transformer-CNN architecture for environmental microorganism classification.一种用于环境微生物分类的新型混合变压器-CNN 架构。
PLoS One. 2022 Nov 11;17(11):e0277557. doi: 10.1371/journal.pone.0277557. eCollection 2022.
8
Dynamic Spatial Sparsification for Efficient Vision Transformers and Convolutional Neural Networks.用于高效视觉Transformer和卷积神经网络的动态空间稀疏化
IEEE Trans Pattern Anal Mach Intell. 2023 Sep;45(9):10883-10897. doi: 10.1109/TPAMI.2023.3263826. Epub 2023 Aug 7.
9
CCTCOVID: COVID-19 detection from chest X-ray images using Compact Convolutional Transformers.CCTCOVID:使用紧凑型卷积变压器从胸部 X 光图像中检测 COVID-19。
Front Public Health. 2023 Feb 27;11:1025746. doi: 10.3389/fpubh.2023.1025746. eCollection 2023.
10
A comparative study on deep learning models for text classification of unstructured medical notes with various levels of class imbalance.深度学习模型在不同类别不平衡程度的非结构化医疗记录文本分类中的对比研究。
BMC Med Res Methodol. 2022 Jul 2;22(1):181. doi: 10.1186/s12874-022-01665-y.