• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于瓶颈融合 Transformer 的共享特定特征学习在多模态全切片图像分析中的应用。

Shared-Specific Feature Learning With Bottleneck Fusion Transformer for Multi-Modal Whole Slide Image Analysis.

出版信息

IEEE Trans Med Imaging. 2023 Nov;42(11):3374-3383. doi: 10.1109/TMI.2023.3287256. Epub 2023 Oct 27.

DOI:10.1109/TMI.2023.3287256
PMID:37335798
Abstract

The fusion of multi-modal medical data is essential to assist medical experts to make treatment decisions for precision medicine. For example, combining the whole slide histopathological images (WSIs) and tabular clinical data can more accurately predict the lymph node metastasis (LNM) of papillary thyroid carcinoma before surgery to avoid unnecessary lymph node resection. However, the huge-sized WSI provides much more high-dimensional information than low-dimensional tabular clinical data, making the information alignment challenging in the multi-modal WSI analysis tasks. This paper presents a novel transformer-guided multi-modal multi-instance learning framework to predict lymph node metastasis from both WSIs and tabular clinical data. We first propose an effective multi-instance grouping scheme, named siamese attention-based feature grouping (SAG), to group high-dimensional WSIs into representative low-dimensional feature embeddings for fusion. We then design a novel bottleneck shared-specific feature transfer module (BSFT) to explore the shared and specific features between different modalities, where a few learnable bottleneck tokens are utilized for knowledge transfer between modalities. Moreover, a modal adaptation and orthogonal projection scheme were incorporated to further encourage BSFT to learn shared and specific features from multi-modal data. Finally, the shared and specific features are dynamically aggregated via an attention mechanism for slide-level prediction. Experimental results on our collected lymph node metastasis dataset demonstrate the efficiency of our proposed components and our framework achieves the best performance with AUC (area under the curve) of 97.34%, outperforming the state-of-the-art methods by over 1.27%.

摘要

多模态医学数据的融合对于辅助医学专家做出精准医疗治疗决策至关重要。例如,将全切片组织病理学图像(WSI)和表格临床数据相结合,可以更准确地预测甲状腺乳头状癌手术前的淋巴结转移(LNM),从而避免不必要的淋巴结切除。然而,巨大尺寸的 WSI 提供的高维信息远多于低维表格临床数据,这使得在多模态 WSI 分析任务中进行信息对齐极具挑战性。本文提出了一种新颖的基于 Transformer 的多模态多实例学习框架,用于从 WSI 和表格临床数据预测淋巴结转移。我们首先提出了一种有效的多实例分组方案,称为孪生注意力特征分组(SAG),用于将高维 WSI 分组为具有代表性的低维特征嵌入以进行融合。然后,我们设计了一种新颖的瓶颈共享特定特征迁移模块(BSFT),用于探索不同模态之间的共享和特定特征,其中利用了几个可学习的瓶颈令牌来实现模态之间的知识迁移。此外,我们还采用了模态自适应和正交投影方案,进一步鼓励 BSFT 从多模态数据中学习共享和特定特征。最后,通过注意力机制对幻灯片级别的预测进行动态聚合共享和特定特征。在我们收集的淋巴结转移数据集上的实验结果表明,我们提出的组件的效率很高,我们的框架的 AUC(曲线下面积)达到 97.34%,性能优于最先进的方法超过 1.27%。

相似文献

1
Shared-Specific Feature Learning With Bottleneck Fusion Transformer for Multi-Modal Whole Slide Image Analysis.基于瓶颈融合 Transformer 的共享特定特征学习在多模态全切片图像分析中的应用。
IEEE Trans Med Imaging. 2023 Nov;42(11):3374-3383. doi: 10.1109/TMI.2023.3287256. Epub 2023 Oct 27.
2
Lymph Node Metastasis Prediction From Whole Slide Images With Transformer-Guided Multiinstance Learning and Knowledge Transfer.基于 Transformer 引导的多实例学习和知识迁移的全切片图像淋巴结转移预测。
IEEE Trans Med Imaging. 2022 Oct;41(10):2777-2787. doi: 10.1109/TMI.2022.3171418. Epub 2022 Sep 30.
3
A modality-collaborative convolution and transformer hybrid network for unpaired multi-modal medical image segmentation with limited annotations.一种用于具有有限标注的未配对多模态医学图像分割的模态协作卷积与Transformer混合网络。
Med Phys. 2023 Sep;50(9):5460-5478. doi: 10.1002/mp.16338. Epub 2023 Mar 15.
4
Second-order multi-instance learning model for whole slide image classification.用于全切片图像分类的二阶多实例学习模型。
Phys Med Biol. 2021 Jul 12;66(14). doi: 10.1088/1361-6560/ac0f30.
5
Colorectal cancer lymph node metastasis prediction with weakly supervised transformer-based multi-instance learning.基于弱监督转换器的多实例学习预测结直肠癌淋巴结转移。
Med Biol Eng Comput. 2023 Jun;61(6):1565-1580. doi: 10.1007/s11517-023-02799-x. Epub 2023 Feb 21.
6
Identification of lymph node metastasis in pre-operation cervical cancer patients by weakly supervised deep learning from histopathological whole-slide biopsy images.基于组织病理全切片图像的弱监督深度学习在术前宫颈癌患者淋巴结转移中的识别。
Cancer Med. 2023 Sep;12(17):17952-17966. doi: 10.1002/cam4.6437. Epub 2023 Aug 10.
7
Automated multi-modal Transformer network (AMTNet) for 3D medical images segmentation.用于3D医学图像分割的自动多模态Transformer网络(AMTNet)。
Phys Med Biol. 2023 Jan 9;68(2). doi: 10.1088/1361-6560/aca74c.
8
Multi-Scale Efficient Graph-Transformer for Whole Slide Image Classification.多尺度高效图Transformer 用于全幻灯片图像分类。
IEEE J Biomed Health Inform. 2023 Dec;27(12):5926-5936. doi: 10.1109/JBHI.2023.3317067. Epub 2023 Dec 5.
9
MC-ViT: Multi-path cross-scale vision transformer for thymoma histopathology whole slide image typing.MC-ViT:用于胸腺瘤组织病理学全切片图像分型的多路径跨尺度视觉Transformer
Front Oncol. 2022 Oct 31;12:925903. doi: 10.3389/fonc.2022.925903. eCollection 2022.
10
Masked autoencoders with handcrafted feature predictions: Transformer for weakly supervised esophageal cancer classification.基于手工特征预测的掩码自动编码器:用于弱监督食管癌分类的 Transformer。
Comput Methods Programs Biomed. 2024 Feb;244:107936. doi: 10.1016/j.cmpb.2023.107936. Epub 2023 Nov 22.

引用本文的文献

1
Large-vocabulary forensic pathological analyses via prototypical cross-modal contrastive learning.通过原型跨模态对比学习进行大词汇量法医病理学分析
Nat Commun. 2025 Jul 23;16(1):6773. doi: 10.1038/s41467-025-62060-x.
2
MSFT-transformer: a multistage fusion tabular transformer for disease prediction using metagenomic data.微软变压器:一种用于使用宏基因组数据进行疾病预测的多级融合表格变压器。
Brief Bioinform. 2025 May 1;26(3). doi: 10.1093/bib/bbaf217.
3
A survey of Transformer applications for histopathological image analysis: New developments and future directions.
基于 Transformer 的病理图像分析应用综述:新进展与未来方向。
Biomed Eng Online. 2023 Sep 25;22(1):96. doi: 10.1186/s12938-023-01157-0.