• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

SAFFusion:一种用于多模态医学图像融合的显著感知频率融合网络。

SAFFusion: a saliency-aware frequency fusion network for multimodal medical image fusion.

作者信息

Liu Renhe, Liu Yu, Wang Han, Li Junxian, Hu Kai

机构信息

School of Microelectronics, Tianjin University, Tianjin 300072, China.

Tianjin Medical University Cancer Institute & Hospital, Tianjin 300060, China.

出版信息

Biomed Opt Express. 2025 May 27;16(6):2459-2481. doi: 10.1364/BOE.555458. eCollection 2025 Jun 1.

DOI:10.1364/BOE.555458
PMID:40677382
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12265500/
Abstract

Medical image fusion integrates complementary information from multimodal medical images to provide comprehensive references for clinical decision-making, such as the diagnosis of Alzheimer's disease and the detection and segmentation of brain tumors. Although traditional and deep learning-based fusion methods have been extensively studied, they often fail to devise targeted strategies that fully utilize distinct regional or feature-specific information. This paper proposes SAFFusion, a saliency-aware frequency fusion network that integrates intensity and texture cues from multimodal medical images. We first introduce Mamba-UNet, a multiscale encoder-decoder architecture enhanced by the Mamba design, to improve global modeling in feature extraction and image reconstruction. By employing the contourlet transform in Mamba-UNet, we replace conventional pooling with multiscale representations and decompose spatial features into high- and low-frequency subbands. A dual-branch frequency feature fusion module then fuses cross-modality information according to the distinct characteristics of these frequency subbands. Furthermore, we apply latent low-rank representation (LatLRR) to assess image saliency and implement adaptive loss constraints to preserve information in salient and non-salient regions. Quantitative results on CT/MRI, SPECT/MRI, and PET/MRI fusion tasks show that SAFFusion outperforms state-of-the-art methods. Qualitative evaluations confirm that SAFFusion effectively merges prominent intensity features and rich textures from multiple sources.

摘要

医学图像融合整合来自多模态医学图像的互补信息,为临床决策提供全面参考,如阿尔茨海默病的诊断以及脑肿瘤的检测与分割。尽管传统的和基于深度学习的融合方法已得到广泛研究,但它们往往未能设计出充分利用不同区域或特定特征信息的针对性策略。本文提出了SAFFusion,一种显著感知频率融合网络,它整合了多模态医学图像的强度和纹理线索。我们首先引入Mamba-UNet,一种由Mamba设计增强的多尺度编码器-解码器架构,以改进特征提取和图像重建中的全局建模。通过在Mamba-UNet中采用轮廓波变换,我们用多尺度表示取代传统池化,并将空间特征分解为高频和低频子带。然后,一个双分支频率特征融合模块根据这些频率子带的不同特征融合跨模态信息。此外,我们应用潜在低秩表示(LatLRR)来评估图像显著性,并实施自适应损失约束以保留显著和非显著区域的信息。在CT/MRI、SPECT/MRI和PET/MRI融合任务上的定量结果表明,SAFFusion优于现有方法。定性评估证实,SAFFusion有效地融合了来自多个源的突出强度特征和丰富纹理。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2b9c/12265500/a92e67765ef3/boe-16-6-2459-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2b9c/12265500/267e3cfad919/boe-16-6-2459-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2b9c/12265500/f9dade54d7d2/boe-16-6-2459-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2b9c/12265500/e79dd4e44a34/boe-16-6-2459-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2b9c/12265500/0301c7ce2aee/boe-16-6-2459-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2b9c/12265500/b72daa1be630/boe-16-6-2459-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2b9c/12265500/9bba0d776883/boe-16-6-2459-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2b9c/12265500/b0f36ea04889/boe-16-6-2459-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2b9c/12265500/5c529b6fd975/boe-16-6-2459-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2b9c/12265500/e9bae2ea8f54/boe-16-6-2459-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2b9c/12265500/a92e67765ef3/boe-16-6-2459-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2b9c/12265500/267e3cfad919/boe-16-6-2459-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2b9c/12265500/f9dade54d7d2/boe-16-6-2459-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2b9c/12265500/e79dd4e44a34/boe-16-6-2459-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2b9c/12265500/0301c7ce2aee/boe-16-6-2459-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2b9c/12265500/b72daa1be630/boe-16-6-2459-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2b9c/12265500/9bba0d776883/boe-16-6-2459-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2b9c/12265500/b0f36ea04889/boe-16-6-2459-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2b9c/12265500/5c529b6fd975/boe-16-6-2459-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2b9c/12265500/e9bae2ea8f54/boe-16-6-2459-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2b9c/12265500/a92e67765ef3/boe-16-6-2459-g010.jpg

相似文献

1
SAFFusion: a saliency-aware frequency fusion network for multimodal medical image fusion.SAFFusion:一种用于多模态医学图像融合的显著感知频率融合网络。
Biomed Opt Express. 2025 May 27;16(6):2459-2481. doi: 10.1364/BOE.555458. eCollection 2025 Jun 1.
2
Structural semantic-guided MR synthesis from PET images via a dual cross-attention mechanism.通过双交叉注意力机制从PET图像进行结构语义引导的MR合成。
Med Phys. 2025 Jul;52(7):e17957. doi: 10.1002/mp.17957.
3
Influence of early through late fusion on pancreas segmentation from imperfectly registered multimodal magnetic resonance imaging.早期至晚期融合对来自配准不完善的多模态磁共振成像的胰腺分割的影响。
J Med Imaging (Bellingham). 2025 Mar;12(2):024008. doi: 10.1117/1.JMI.12.2.024008. Epub 2025 Apr 26.
4
Noise-aware system generative model (NASGM): positron emission tomography (PET) image simulation framework with observer validation studies.噪声感知系统生成模型(NASGM):用于正电子发射断层扫描(PET)图像模拟框架及观察者验证研究。
Med Phys. 2025 Jul;52(7):e17962. doi: 10.1002/mp.17962.
5
TLTNet: A novel transscale cascade layered transformer network for enhanced retinal blood vessel segmentation.TLTNet:一种新颖的跨尺度级联分层Transformer 网络,用于增强视网膜血管分割。
Comput Biol Med. 2024 Aug;178:108773. doi: 10.1016/j.compbiomed.2024.108773. Epub 2024 Jun 25.
6
Short-Term Memory Impairment短期记忆障碍
7
VMDU-net: a dual encoder multi-scale fusion network for polyp segmentation with Vision Mamba and Cross-Shape Transformer integration.VMDU-net:一种用于息肉分割的双编码器多尺度融合网络,集成了视觉曼巴和十字形变换器
Front Artif Intell. 2025 Jun 18;8:1557508. doi: 10.3389/frai.2025.1557508. eCollection 2025.
8
DCMC-UNet: A Novel Segmentation Model for Carbon Traces in Oil-Immersed Transformers Improved with Dynamic Feature Fusion and Adaptive Illumination Enhancement.DCMC-UNet:一种通过动态特征融合和自适应光照增强改进的油浸式变压器碳痕分割新模型。
Sensors (Basel). 2025 Jun 23;25(13):3904. doi: 10.3390/s25133904.
9
A novel recursive transformer-based U-Net architecture for enhanced multi-scale medical image segmentation.一种基于递归变压器的新型U-Net架构,用于增强多尺度医学图像分割。
Comput Biol Med. 2025 Sep;196(Pt A):110658. doi: 10.1016/j.compbiomed.2025.110658. Epub 2025 Jul 6.
10
SODU2-NET: a novel deep learning-based approach for salient object detection utilizing U-NET.SODU2-NET:一种基于深度学习的利用U-NET进行显著目标检测的新方法。
PeerJ Comput Sci. 2025 May 19;11:e2623. doi: 10.7717/peerj-cs.2623. eCollection 2025.

本文引用的文献

1
Multi-Dimensional Medical Image Fusion With Complex Sparse Representation.基于复杂稀疏表示的多维医学图像融合
IEEE Trans Biomed Eng. 2024 Sep;71(9):2728-2739. doi: 10.1109/TBME.2024.3391314. Epub 2024 Aug 21.
2
TCGAN: a transformer-enhanced GAN for PET synthetic CT.TCGAN:一种用于PET合成CT的变压器增强型生成对抗网络。
Biomed Opt Express. 2022 Oct 24;13(11):6003-6018. doi: 10.1364/BOE.467683. eCollection 2022 Nov 1.
3
ERA-WGAT: Edge-enhanced residual autoencoder with a window-based graph attention convolutional network for low-dose CT denoising.
ERA-WGAT:用于低剂量CT去噪的基于窗口的图注意力卷积网络的边缘增强残差自动编码器
Biomed Opt Express. 2022 Oct 13;13(11):5775-5793. doi: 10.1364/BOE.471340. eCollection 2022 Nov 1.
4
MATR: Multimodal Medical Image Fusion via Multiscale Adaptive Transformer.MATR:基于多尺度自适应变换的多模态医学图像融合。
IEEE Trans Image Process. 2022;31:5134-5149. doi: 10.1109/TIP.2022.3193288. Epub 2022 Aug 2.
5
Learning Enriched Features for Fast Image Restoration and Enhancement.学习丰富特征以实现快速图像恢复与增强。
IEEE Trans Pattern Anal Mach Intell. 2023 Feb;45(2):1934-1948. doi: 10.1109/TPAMI.2022.3167175. Epub 2023 Jan 6.
6
U2Fusion: A Unified Unsupervised Image Fusion Network.U2Fusion:一种统一的无监督图像融合网络。
IEEE Trans Pattern Anal Mach Intell. 2022 Jan;44(1):502-518. doi: 10.1109/TPAMI.2020.3012548. Epub 2021 Dec 8.
7
DDcGAN: A Dual-discriminator Conditional Generative Adversarial Network for Multi-resolution Image Fusion.DDcGAN:一种用于多分辨率图像融合的双判别器条件生成对抗网络。
IEEE Trans Image Process. 2020 Mar 10. doi: 10.1109/TIP.2020.2977573.
8
MDLatLRR: A novel decomposition method for infrared and visible image fusion.MDLatLRR:一种用于红外与可见光图像融合的新型分解方法。
IEEE Trans Image Process. 2020 Feb 28. doi: 10.1109/TIP.2020.2975984.
9
Medical Image Fusion Based on Fast Finite Shearlet Transform and Sparse Representation.基于快速有限剪切波变换和稀疏表示的医学图像融合
Comput Math Methods Med. 2019 Mar 3;2019:3503267. doi: 10.1155/2019/3503267. eCollection 2019.
10
Multimodal neuroimaging feature learning for multiclass diagnosis of Alzheimer's disease.用于阿尔茨海默病多类诊断的多模态神经影像特征学习
IEEE Trans Biomed Eng. 2015 Apr;62(4):1132-40. doi: 10.1109/TBME.2014.2372011. Epub 2014 Nov 20.