• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于Swin变压器的生成对抗网络用于多模态医学图像翻译。

Swin transformer-based GAN for multi-modal medical image translation.

作者信息

Yan Shouang, Wang Chengyan, Chen Weibo, Lyu Jun

机构信息

School of Computer and Control Engineering, Yantai University, Yantai, China.

Human Phenome Institute, Fudan University, Shanghai, China.

出版信息

Front Oncol. 2022 Aug 8;12:942511. doi: 10.3389/fonc.2022.942511. eCollection 2022.

DOI:10.3389/fonc.2022.942511
PMID:36003791
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC9395186/
Abstract

Medical image-to-image translation is considered a new direction with many potential applications in the medical field. The medical image-to-image translation is dominated by two models, including supervised Pix2Pix and unsupervised cyclic-consistency generative adversarial network (GAN). However, existing methods still have two shortcomings: 1) the Pix2Pix requires paired and pixel-aligned images, which are difficult to acquire. Nevertheless, the optimum output of the cycle-consistency model may not be unique. 2) They are still deficient in capturing the global features and modeling long-distance interactions, which are critical for regions with complex anatomical structures. We propose a Swin Transformer-based GAN for Multi-Modal Medical Image Translation, named MMTrans. Specifically, MMTrans consists of a generator, a registration network, and a discriminator. The Swin Transformer-based generator enables to generate images with the same content as source modality images and similar style information of target modality images. The encoder part of the registration network, based on Swin Transformer, is utilized to predict deformable vector fields. The convolution-based discriminator determines whether the target modality images are similar to the generator or from the real images. Extensive experiments conducted using the public dataset and clinical datasets showed that our network outperformed other advanced medical image translation methods in both aligned and unpaired datasets and has great potential to be applied in clinical applications.

摘要

医学图像到图像的转换被认为是一个新的方向,在医学领域有许多潜在应用。医学图像到图像的转换主要由两种模型主导,包括有监督的Pix2Pix和无监督的循环一致生成对抗网络(GAN)。然而,现有方法仍存在两个缺点:1)Pix2Pix需要配对且像素对齐的图像,而这些图像很难获取。此外,循环一致模型的最优输出可能不唯一。2)它们在捕捉全局特征和对长距离交互进行建模方面仍然存在不足,而这对于具有复杂解剖结构的区域至关重要。我们提出了一种基于Swin Transformer的用于多模态医学图像转换的GAN,名为MMTrans。具体而言,MMTrans由一个生成器、一个配准网络和一个判别器组成。基于Swin Transformer的生成器能够生成与源模态图像具有相同内容且与目标模态图像具有相似风格信息的图像。基于Swin Transformer的配准网络的编码器部分用于预测可变形向量场。基于卷积的判别器确定目标模态图像是与生成器生成的图像相似还是来自真实图像。使用公共数据集和临床数据集进行的大量实验表明,我们的网络在对齐和未配对数据集中均优于其他先进的医学图像转换方法,并且在临床应用中具有很大的应用潜力。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cd9a/9395186/196aff0cd507/fonc-12-942511-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cd9a/9395186/7bbaf0c82b03/fonc-12-942511-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cd9a/9395186/e5b1e483cf7c/fonc-12-942511-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cd9a/9395186/5297329dd738/fonc-12-942511-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cd9a/9395186/9b04802cbfa8/fonc-12-942511-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cd9a/9395186/2ddf1d4850d5/fonc-12-942511-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cd9a/9395186/d41e46e2f413/fonc-12-942511-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cd9a/9395186/57a3cca81a33/fonc-12-942511-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cd9a/9395186/ad12c6cd0e8b/fonc-12-942511-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cd9a/9395186/6728a1263439/fonc-12-942511-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cd9a/9395186/2121faddaaee/fonc-12-942511-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cd9a/9395186/196aff0cd507/fonc-12-942511-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cd9a/9395186/7bbaf0c82b03/fonc-12-942511-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cd9a/9395186/e5b1e483cf7c/fonc-12-942511-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cd9a/9395186/5297329dd738/fonc-12-942511-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cd9a/9395186/9b04802cbfa8/fonc-12-942511-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cd9a/9395186/2ddf1d4850d5/fonc-12-942511-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cd9a/9395186/d41e46e2f413/fonc-12-942511-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cd9a/9395186/57a3cca81a33/fonc-12-942511-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cd9a/9395186/ad12c6cd0e8b/fonc-12-942511-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cd9a/9395186/6728a1263439/fonc-12-942511-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cd9a/9395186/2121faddaaee/fonc-12-942511-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cd9a/9395186/196aff0cd507/fonc-12-942511-g011.jpg

相似文献

1
Swin transformer-based GAN for multi-modal medical image translation.基于Swin变压器的生成对抗网络用于多模态医学图像翻译。
Front Oncol. 2022 Aug 8;12:942511. doi: 10.3389/fonc.2022.942511. eCollection 2022.
2
A modality-collaborative convolution and transformer hybrid network for unpaired multi-modal medical image segmentation with limited annotations.一种用于具有有限标注的未配对多模态医学图像分割的模态协作卷积与Transformer混合网络。
Med Phys. 2023 Sep;50(9):5460-5478. doi: 10.1002/mp.16338. Epub 2023 Mar 15.
3
SwinGAN: A dual-domain Swin Transformer-based generative adversarial network for MRI reconstruction.SwinGAN:一种基于双域 Swin Transformer 的生成对抗网络,用于 MRI 重建。
Comput Biol Med. 2023 Feb;153:106513. doi: 10.1016/j.compbiomed.2022.106513. Epub 2022 Dec 31.
4
SwinCross: Cross-modal Swin transformer for head-and-neck tumor segmentation in PET/CT images.SwinCross:用于 PET/CT 图像中头颈部肿瘤分割的跨模态 Swin 变换器。
Med Phys. 2024 Mar;51(3):2096-2107. doi: 10.1002/mp.16703. Epub 2023 Sep 30.
5
ST-GAN: A Swin Transformer-Based Generative Adversarial Network for Unsupervised Domain Adaptation of Cross-Modality Cardiac Segmentation.ST-GAN:一种基于 Swin Transformer 的生成对抗网络,用于跨模态心脏分割的无监督领域自适应。
IEEE J Biomed Health Inform. 2024 Feb;28(2):893-904. doi: 10.1109/JBHI.2023.3336965. Epub 2024 Feb 5.
6
Reliable multi-modal medical image-to-image translation independent of pixel-wise aligned data.可靠的多模态医学图像到图像的转换,不依赖于像素对齐的数据。
Med Phys. 2024 Nov;51(11):8283-8301. doi: 10.1002/mp.17362. Epub 2024 Aug 17.
7
Multi-Modal Brain Tumor Data Completion Based on Reconstruction Consistency Loss.基于重建一致性损失的多模态脑肿瘤数据补全。
J Digit Imaging. 2023 Aug;36(4):1794-1807. doi: 10.1007/s10278-022-00697-6. Epub 2023 Mar 1.
8
Swin-Net: A Swin-Transformer-Based Network Combing with Multi-Scale Features for Segmentation of Breast Tumor Ultrasound Images.Swin-Net:一种基于Swin-Transformer并结合多尺度特征的用于乳腺肿瘤超声图像分割的网络。
Diagnostics (Basel). 2024 Jan 26;14(3):269. doi: 10.3390/diagnostics14030269.
9
BTMF-GAN: A multi-modal MRI fusion generative adversarial network for brain tumors.BTMF-GAN:一种用于脑肿瘤的多模态 MRI 融合生成对抗网络。
Comput Biol Med. 2023 May;157:106769. doi: 10.1016/j.compbiomed.2023.106769. Epub 2023 Mar 9.
10
Swin-MFA: A Multi-Modal Fusion Attention Network Based on Swin-Transformer for Low-Light Image Human Segmentation.Swin-MFA:一种基于 Swin-Transformer 的多模态融合注意力网络,用于低光照图像人体分割。
Sensors (Basel). 2022 Aug 19;22(16):6229. doi: 10.3390/s22166229.

引用本文的文献

1
Robust deep MRI contrast synthesis using a prior-based and task-oriented 3D network.使用基于先验和面向任务的3D网络进行稳健的深度MRI对比度合成。
Imaging Neurosci (Camb). 2025 Aug 26;3. doi: 10.1162/IMAG.a.116. eCollection 2025.
2
Deep Learning in Digital Breast Tomosynthesis: Current Status, Challenges, and Future Trends.数字乳腺断层合成中的深度学习:现状、挑战与未来趋势。
MedComm (2020). 2025 Jun 9;6(6):e70247. doi: 10.1002/mco2.70247. eCollection 2025 Jun.
3
Based on TransRes-Pix2Pix network to generate the OBL image during SMILE surgery.

本文引用的文献

1
ReconFormer: Accelerated MRI Reconstruction Using Recurrent Transformer.ReconFormer:基于循环 Transformer 的加速 MRI 重建。
IEEE Trans Med Imaging. 2024 Jan;43(1):582-593. doi: 10.1109/TMI.2023.3314747. Epub 2024 Jan 2.
2
Cross-Modality LGE-CMR Segmentation Using Image-to-Image Translation Based Data Augmentation.基于图像到图像翻译的数据增强的跨模态LGE-CMR分割
IEEE/ACM Trans Comput Biol Bioinform. 2023 Jul-Aug;20(4):2367-2375. doi: 10.1109/TCBB.2022.3140306. Epub 2023 Aug 9.
3
Multi-Modal MRI Image Synthesis via GAN With Multi-Scale Gate Mergence.
基于TransRes-Pix2Pix网络在微笑手术期间生成口腔颌面部图像。
Front Cell Dev Biol. 2025 May 21;13:1598475. doi: 10.3389/fcell.2025.1598475. eCollection 2025.
4
The Role of Artificial Intelligence in Epiretinal Membrane Care: A Scoping Review.人工智能在视网膜前膜护理中的作用:一项范围综述
Ophthalmol Sci. 2024 Dec 20;5(4):100689. doi: 10.1016/j.xops.2024.100689. eCollection 2025 Jul-Aug.
5
A multi-modal deep learning solution for precise pneumonia diagnosis: the PneumoFusion-Net model.一种用于精确肺炎诊断的多模态深度学习解决方案:PneumoFusion-Net模型。
Front Physiol. 2025 Mar 12;16:1512835. doi: 10.3389/fphys.2025.1512835. eCollection 2025.
6
Transformers and large language models in healthcare: A review.医疗保健中的变压器和大型语言模型:综述。
Artif Intell Med. 2024 Aug;154:102900. doi: 10.1016/j.artmed.2024.102900. Epub 2024 Jun 5.
7
L2NLF: a novel linear-to-nonlinear framework for multi-modal medical image registration.L2NLF:一种用于多模态医学图像配准的新型线性到非线性框架。
Biomed Eng Lett. 2024 Jan 10;14(3):497-509. doi: 10.1007/s13534-023-00344-1. eCollection 2024 May.
8
Exceeding the limit for microscopic image translation with a deep learning-based unified framework.基于深度学习的统一框架突破微观图像翻译的极限。
PNAS Nexus. 2024 Mar 29;3(4):pgae133. doi: 10.1093/pnasnexus/pgae133. eCollection 2024 Apr.
9
A Deep Diagnostic Framework Using Explainable Artificial Intelligence and Clustering.一种使用可解释人工智能和聚类的深度诊断框架。
Diagnostics (Basel). 2023 Nov 9;13(22):3413. doi: 10.3390/diagnostics13223413.
10
SW-UNet: a U-Net fusing sliding window transformer block with CNN for segmentation of lung nodules.SW-UNet:一种将滑动窗口变压器模块与卷积神经网络融合用于肺结节分割的U-Net。
Front Med (Lausanne). 2023 Sep 28;10:1273441. doi: 10.3389/fmed.2023.1273441. eCollection 2023.
基于多尺度门控融合的生成对抗网络的多模态 MRI 图像合成。
IEEE J Biomed Health Inform. 2022 Jan;26(1):17-26. doi: 10.1109/JBHI.2021.3088866. Epub 2022 Jan 17.
4
Deep learning-based multi-modal computing with feature disentanglement for MRI image synthesis.基于深度学习的多模态计算与特征解缠用于 MRI 图像合成。
Med Phys. 2021 Jul;48(7):3778-3789. doi: 10.1002/mp.14929. Epub 2021 Jun 7.
5
Adversarial Uni- and Multi-modal Stream Networks for Multimodal Image Registration.用于多模态图像配准的对抗式单模态和多模态流网络
Med Image Comput Comput Assist Interv. 2020 Oct;12263:222-232. doi: 10.1007/978-3-030-59716-0_22. Epub 2020 Sep 29.
6
Hi-Net: Hybrid-Fusion Network for Multi-Modal MR Image Synthesis.Hi-Net:用于多模态磁共振图像合成的混合融合网络。
IEEE Trans Med Imaging. 2020 Sep;39(9):2772-2781. doi: 10.1109/TMI.2020.2975344. Epub 2020 Feb 20.
7
Image Synthesis in Multi-Contrast MRI With Conditional Generative Adversarial Networks.多对比度 MRI 图像合成的条件生成对抗网络。
IEEE Trans Med Imaging. 2019 Oct;38(10):2375-2388. doi: 10.1109/TMI.2019.2901750. Epub 2019 Feb 26.
8
3D Auto-Context-Based Locality Adaptive Multi-Modality GANs for PET Synthesis.基于 3D 自动上下文的局部自适应多模态 GAN 用于 PET 合成。
IEEE Trans Med Imaging. 2019 Jun;38(6):1328-1339. doi: 10.1109/TMI.2018.2884053. Epub 2018 Nov 29.
9
The Multimodal Brain Tumor Image Segmentation Benchmark (BRATS).多模态脑肿瘤图像分割基准(BRATS)。
IEEE Trans Med Imaging. 2015 Oct;34(10):1993-2024. doi: 10.1109/TMI.2014.2377694. Epub 2014 Dec 4.
10
Multi-contrast, isotropic, single-slab 3D MR imaging in multiple sclerosis.多发性硬化症的多对比、各向同性、单层3D磁共振成像
Eur Radiol. 2008 Oct;18(10):2311-20. doi: 10.1007/s00330-008-1009-7. Epub 2008 May 29.