• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

MATR:基于多尺度自适应变换的多模态医学图像融合。

MATR: Multimodal Medical Image Fusion via Multiscale Adaptive Transformer.

出版信息

IEEE Trans Image Process. 2022;31:5134-5149. doi: 10.1109/TIP.2022.3193288. Epub 2022 Aug 2.

DOI:10.1109/TIP.2022.3193288
PMID:35901003
Abstract

Owing to the limitations of imaging sensors, it is challenging to obtain a medical image that simultaneously contains functional metabolic information and structural tissue details. Multimodal medical image fusion, an effective way to merge the complementary information in different modalities, has become a significant technique to facilitate clinical diagnosis and surgical navigation. With powerful feature representation ability, deep learning (DL)-based methods have improved such fusion results but still have not achieved satisfactory performance. Specifically, existing DL-based methods generally depend on convolutional operations, which can well extract local patterns but have limited capability in preserving global context information. To compensate for this defect and achieve accurate fusion, we propose a novel unsupervised method to fuse multimodal medical images via a multiscale adaptive Transformer termed MATR. In the proposed method, instead of directly employing vanilla convolution, we introduce an adaptive convolution for adaptively modulating the convolutional kernel based on the global complementary context. To further model long-range dependencies, an adaptive Transformer is employed to enhance the global semantic extraction capability. Our network architecture is designed in a multiscale fashion so that useful multimodal information can be adequately acquired from the perspective of different scales. Moreover, an objective function composed of a structural loss and a region mutual information loss is devised to construct constraints for information preservation at both the structural-level and the feature-level. Extensive experiments on a mainstream database demonstrate that the proposed method outperforms other representative and state-of-the-art methods in terms of both visual quality and quantitative evaluation. We also extend the proposed method to address other biomedical image fusion issues, and the pleasing fusion results illustrate that MATR has good generalization capability. The code of the proposed method is available at https://github.com/tthinking/MATR.

摘要

由于成像传感器的限制,很难获得同时包含功能代谢信息和结构组织细节的医学图像。多模态医学图像融合是一种融合不同模态互补信息的有效方法,已成为促进临床诊断和手术导航的重要技术。基于深度学习(DL)的方法具有强大的特征表示能力,提高了这种融合效果,但仍未达到令人满意的性能。具体来说,现有的基于 DL 的方法通常依赖于卷积操作,虽然卷积操作可以很好地提取局部模式,但在保留全局上下文信息方面能力有限。为了弥补这一缺陷并实现准确的融合,我们提出了一种新颖的基于多尺度自适应 Transformer 的无监督方法,称为 MATR,用于融合多模态医学图像。在提出的方法中,我们不是直接使用常规卷积,而是引入了自适应卷积,根据全局互补上下文自适应地调整卷积核。为了进一步建模远程依赖关系,引入自适应 Transformer 来增强全局语义提取能力。我们的网络架构采用多尺度设计,以便从不同尺度充分获取有用的多模态信息。此外,我们设计了一个由结构损失和区域互信息损失组成的目标函数,以在结构级和特征级构建信息保留的约束。在主流数据库上的广泛实验表明,与其他代表性和最先进的方法相比,所提出的方法在视觉质量和定量评估方面都表现出色。我们还将所提出的方法扩展到解决其他生物医学图像融合问题,令人满意的融合结果表明 MATR 具有良好的泛化能力。该方法的代码可在 https://github.com/tthinking/MATR 上获得。

相似文献

1
MATR: Multimodal Medical Image Fusion via Multiscale Adaptive Transformer.MATR:基于多尺度自适应变换的多模态医学图像融合。
IEEE Trans Image Process. 2022;31:5134-5149. doi: 10.1109/TIP.2022.3193288. Epub 2022 Aug 2.
2
EMOST: A dual-branch hybrid network for medical image fusion via efficient model module and sparse transformer.EMOST:一种基于高效模型模块和稀疏 Transformer 的医学图像融合双分支混合网络。
Comput Biol Med. 2024 Sep;179:108771. doi: 10.1016/j.compbiomed.2024.108771. Epub 2024 Jul 5.
3
A multibranch and multiscale neural network based on semantic perception for multimodal medical image fusion.基于语义感知的多分支多尺度神经网络用于多模态医学图像融合。
Sci Rep. 2024 Jul 30;14(1):17609. doi: 10.1038/s41598-024-68183-3.
4
Spatial adaptive and transformer fusion network (STFNet) for low-count PET blind denoising with MRI.基于 MRI 的低计数 PET 盲去噪的空间自适应和变换融合网络(STFNet)
Med Phys. 2022 Jan;49(1):343-356. doi: 10.1002/mp.15368. Epub 2021 Dec 10.
5
MFNet: Multimodal medical image fusion network via multi-receptive-field and multi-scale feature integration.MFNet:一种基于多感受野和多尺度特征融合的多模态医学图像融合网络。
Comput Biol Med. 2023 Jun;159:106923. doi: 10.1016/j.compbiomed.2023.106923. Epub 2023 Apr 14.
6
MACTFusion: Lightweight Cross Transformer for Adaptive Multimodal Medical Image Fusion.MACTFusion:用于自适应多模态医学图像融合的轻量级交叉变换器
IEEE J Biomed Health Inform. 2025 May;29(5):3317-3328. doi: 10.1109/JBHI.2024.3391620. Epub 2025 May 6.
7
VANet: a medical image fusion model based on attention mechanism to assist disease diagnosis.VANet:一种基于注意力机制的医学图像融合模型,用于辅助疾病诊断。
BMC Bioinformatics. 2022 Dec 19;23(1):548. doi: 10.1186/s12859-022-05072-4.
8
MDC-RHT: Multi-Modal Medical Image Fusion via Multi-Dimensional Dynamic Convolution and Residual Hybrid Transformer.MDC-RHT:基于多维动态卷积和残差混合变压器的多模态医学图像融合
Sensors (Basel). 2024 Jun 21;24(13):4056. doi: 10.3390/s24134056.
9
Multimodal medical image fusion based on interval gradients and convolutional neural networks.基于区间梯度和卷积神经网络的多模态医学图像融合。
BMC Med Imaging. 2024 Sep 5;24(1):232. doi: 10.1186/s12880-024-01418-x.
10
3D Multimodal Fusion Network With Disease-Induced Joint Learning for Early Alzheimer's Disease Diagnosis.基于疾病诱导联合学习的 3D 多模态融合网络用于早期阿尔茨海默病诊断。
IEEE Trans Med Imaging. 2024 Sep;43(9):3161-3175. doi: 10.1109/TMI.2024.3386937. Epub 2024 Sep 3.

引用本文的文献

1
Language-Driven Cross-Attention for Visible-Infrared Image Fusion Using CLIP.使用CLIP的语言驱动交叉注意力用于可见光-红外图像融合
Sensors (Basel). 2025 Aug 15;25(16):5083. doi: 10.3390/s25165083.
2
Unsupervised cross-modal biomedical image fusion framework with dual-path detail enhancement and global context awareness.具有双路径细节增强和全局上下文感知的无监督跨模态生物医学图像融合框架
Biomed Opt Express. 2025 Jul 25;16(8):3378-3394. doi: 10.1364/BOE.562137. eCollection 2025 Aug 1.
3
Deep learning assisted non-invasive lymph node burden evaluation and CDK4/6i administration in luminal breast cancer.
深度学习辅助的管腔型乳腺癌非侵入性淋巴结负荷评估及CDK4/6抑制剂给药
iScience. 2025 Jun 7;28(7):112849. doi: 10.1016/j.isci.2025.112849. eCollection 2025 Jul 18.
4
SAFFusion: a saliency-aware frequency fusion network for multimodal medical image fusion.SAFFusion:一种用于多模态医学图像融合的显著感知频率融合网络。
Biomed Opt Express. 2025 May 27;16(6):2459-2481. doi: 10.1364/BOE.555458. eCollection 2025 Jun 1.
5
Saliency-enhanced infrared and visible image fusion via sub-window variance filter and weighted least squares optimization.基于子窗口方差滤波器和加权最小二乘优化的显著性增强红外与可见光图像融合
PLoS One. 2025 Jul 7;20(7):e0323285. doi: 10.1371/journal.pone.0323285. eCollection 2025.
6
A novel multimodal computer-aided diagnostic model for pulmonary embolism based on hybrid transformer-CNN and tabular transformer.一种基于混合Transformer-CNN和表格Transformer的新型多模态肺栓塞计算机辅助诊断模型。
Phys Eng Sci Med. 2025 May 24. doi: 10.1007/s13246-025-01568-4.
7
TongueNet: a multi-modal fusion and multi-label classification model for traditional Chinese Medicine tongue diagnosis.舌诊网络:一种用于中医舌诊的多模态融合与多标签分类模型。
Front Physiol. 2025 Apr 25;16:1527751. doi: 10.3389/fphys.2025.1527751. eCollection 2025.
8
A multi-scale pyramid residual weight network for medical image fusion.一种用于医学图像融合的多尺度金字塔残差权重网络。
Quant Imaging Med Surg. 2025 Mar 3;15(3):1793-1821. doi: 10.21037/qims-24-851. Epub 2025 Feb 26.
9
A stochastic structural similarity guided approach for multi-modal medical image fusion.一种基于随机结构相似性引导的多模态医学图像融合方法。
Sci Rep. 2025 Mar 14;15(1):8792. doi: 10.1038/s41598-025-93662-6.
10
Diffusion-driven multi-modality medical image fusion.扩散驱动的多模态医学图像融合。
Med Biol Eng Comput. 2025 Feb 11. doi: 10.1007/s11517-025-03300-6.