• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

CPFTransformer:变换器融合上下文金字塔医学图像分割网络。

CPFTransformer: transformer fusion context pyramid medical image segmentation network.

作者信息

Li Jiao, Ye Jinyu, Zhang Ruixin, Wu Yue, Berhane Gebremedhin Samuel, Deng Hongxia, Shi Hong

机构信息

College of Computer Science and Technology, Taiyuan University of Technology, Taiyuan, China.

School of Artificial Intelligence, Shenzhen Polytechnic, Shenzhen, China.

出版信息

Front Neurosci. 2023 Dec 7;17:1288366. doi: 10.3389/fnins.2023.1288366. eCollection 2023.

DOI:10.3389/fnins.2023.1288366
PMID:38130692
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10733526/
Abstract

INTRODUCTION

The application of U-shaped convolutional neural network (CNN) methods in medical image segmentation tasks has yielded impressive results. However, this structure's single-level context information extraction capability can lead to problems such as boundary blurring, so it needs to be improved. Additionally, the convolution operation's inherent locality restricts its ability to capture global and long-distance semantic information interactions effectively. Conversely, the transformer model excels at capturing global information.

METHODS

Given these considerations, this paper presents a transformer fusion context pyramid medical image segmentation network (CPFTransformer). The CPFTransformer utilizes the Swin Transformer to integrate edge perception for segmentation edges. To effectively fuse global and multi-scale context information, we introduce an Edge-Aware module based on a context pyramid, which specifically emphasizes local features like edges and corners. Our approach employs a layered Swin Transformer with a shifted window mechanism as an encoder to extract contextual features. A decoder based on a symmetric Swin Transformer is employed for upsampling operations, thereby restoring the resolution of feature maps. The encoder and decoder are connected by an Edge-Aware module for the extraction of local features such as edges and corners.

RESULTS

Experimental evaluations on the Synapse multi-organ segmentation task and the ACDC dataset demonstrate the effectiveness of our method, yielding a segmentation accuracy of 79.87% (DSC) and 20.83% (HD) in the Synapse multi-organ segmentation task.

DISCUSSION

The method proposed in this paper, which combines the context pyramid mechanism and Transformer, enables fast and accurate automatic segmentation of medical images, thereby significantly enhancing the precision and reliability of medical diagnosis. Furthermore, the approach presented in this study can potentially be extended to image segmentation of other organs in the future.

摘要

引言

U型卷积神经网络(CNN)方法在医学图像分割任务中的应用取得了令人瞩目的成果。然而,这种结构的单级上下文信息提取能力可能会导致边界模糊等问题,因此需要改进。此外,卷积操作固有的局部性限制了其有效捕捉全局和长距离语义信息交互的能力。相反,Transformer模型在捕捉全局信息方面表现出色。

方法

考虑到这些因素,本文提出了一种Transformer融合上下文金字塔医学图像分割网络(CPFTransformer)。CPFTransformer利用Swin Transformer来整合分割边缘的边缘感知。为了有效地融合全局和多尺度上下文信息,我们引入了一种基于上下文金字塔的边缘感知模块,该模块特别强调边缘和角落等局部特征。我们的方法采用具有移位窗口机制的分层Swin Transformer作为编码器来提取上下文特征。基于对称Swin Transformer的解码器用于上采样操作,从而恢复特征图的分辨率。编码器和解码器通过边缘感知模块连接,用于提取边缘和角落等局部特征。

结果

在Synapse多器官分割任务和ACDC数据集上的实验评估证明了我们方法的有效性,在Synapse多器官分割任务中分割准确率达到79.87%(DSC)和20.83%(HD)。

讨论

本文提出的方法结合了上下文金字塔机制和Transformer,能够实现医学图像的快速准确自动分割,从而显著提高医学诊断的精度和可靠性。此外,本研究提出的方法未来有可能扩展到其他器官的图像分割。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/25fe/10733526/2a0aeaf677d6/fnins-17-1288366-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/25fe/10733526/fae9880b3432/fnins-17-1288366-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/25fe/10733526/cc2eea297370/fnins-17-1288366-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/25fe/10733526/fe929fdad4ce/fnins-17-1288366-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/25fe/10733526/d557205128d2/fnins-17-1288366-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/25fe/10733526/75d51b05f426/fnins-17-1288366-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/25fe/10733526/2a0aeaf677d6/fnins-17-1288366-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/25fe/10733526/fae9880b3432/fnins-17-1288366-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/25fe/10733526/cc2eea297370/fnins-17-1288366-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/25fe/10733526/fe929fdad4ce/fnins-17-1288366-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/25fe/10733526/d557205128d2/fnins-17-1288366-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/25fe/10733526/75d51b05f426/fnins-17-1288366-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/25fe/10733526/2a0aeaf677d6/fnins-17-1288366-g006.jpg

相似文献

1
CPFTransformer: transformer fusion context pyramid medical image segmentation network.CPFTransformer:变换器融合上下文金字塔医学图像分割网络。
Front Neurosci. 2023 Dec 7;17:1288366. doi: 10.3389/fnins.2023.1288366. eCollection 2023.
2
Dual encoder network with transformer-CNN for multi-organ segmentation.基于 Transformer-CNN 的双编码器网络的多器官分割。
Med Biol Eng Comput. 2023 Mar;61(3):661-671. doi: 10.1007/s11517-022-02723-9. Epub 2022 Dec 29.
3
TLTNet: A novel transscale cascade layered transformer network for enhanced retinal blood vessel segmentation.TLTNet:一种新颖的跨尺度级联分层Transformer 网络,用于增强视网膜血管分割。
Comput Biol Med. 2024 Aug;178:108773. doi: 10.1016/j.compbiomed.2024.108773. Epub 2024 Jun 25.
4
Transformer-Based Model with Dynamic Attention Pyramid Head for Semantic Segmentation of VHR Remote Sensing Imagery.基于Transformer且带有动态注意力金字塔头的甚高分辨率遥感影像语义分割模型
Entropy (Basel). 2022 Nov 6;24(11):1619. doi: 10.3390/e24111619.
5
A new architecture combining convolutional and transformer-based networks for automatic 3D multi-organ segmentation on CT images.一种新的架构,结合了卷积和基于Transformer 的网络,用于 CT 图像上的自动 3D 多器官分割。
Med Phys. 2023 Nov;50(11):6990-7002. doi: 10.1002/mp.16750. Epub 2023 Sep 22.
6
Swin-Net: A Swin-Transformer-Based Network Combing with Multi-Scale Features for Segmentation of Breast Tumor Ultrasound Images.Swin-Net:一种基于Swin-Transformer并结合多尺度特征的用于乳腺肿瘤超声图像分割的网络。
Diagnostics (Basel). 2024 Jan 26;14(3):269. doi: 10.3390/diagnostics14030269.
7
MSCT-UNET: multi-scale contrastive transformer within U-shaped network for medical image segmentation.MSCT-UNET:U 形网络中的多尺度对比变换用于医学图像分割。
Phys Med Biol. 2023 Dec 28;69(1). doi: 10.1088/1361-6560/ad135d.
8
MS-TCNet: An effective Transformer-CNN combined network using multi-scale feature learning for 3D medical image segmentation.MS-TCNet:一种基于多尺度特征学习的有效的 Transformer-CNN 组合网络,用于 3D 医学图像分割。
Comput Biol Med. 2024 Mar;170:108057. doi: 10.1016/j.compbiomed.2024.108057. Epub 2024 Jan 28.
9
SwinBTS: A Method for 3D Multimodal Brain Tumor Segmentation Using Swin Transformer.SwinBTS:一种使用Swin Transformer进行3D多模态脑肿瘤分割的方法。
Brain Sci. 2022 Jun 17;12(6):797. doi: 10.3390/brainsci12060797.
10
STEDNet: Swin transformer-based encoder-decoder network for noise reduction in low-dose CT.STEDNet:基于 Swin Transformer 的编解码网络,用于降低低剂量 CT 中的噪声。
Med Phys. 2023 Jul;50(7):4443-4458. doi: 10.1002/mp.16249. Epub 2023 Feb 9.

引用本文的文献

1
SMoFFI-SegFormer: a novel approach for ovarian tumor segmentation based on an improved SegFormer architecture.SMoFFI-SegFormer:一种基于改进的SegFormer架构的卵巢肿瘤分割新方法。
Front Oncol. 2025 Jul 21;15:1555585. doi: 10.3389/fonc.2025.1555585. eCollection 2025.

本文引用的文献

1
SWTRU: Star-shaped Window Transformer Reinforced U-Net for medical image segmentation.SWTRU:星型窗口变换增强 U-Net 用于医学图像分割。
Comput Biol Med. 2022 Nov;150:105954. doi: 10.1016/j.compbiomed.2022.105954. Epub 2022 Aug 13.
2
Boundary loss for highly unbalanced segmentation.高度不平衡分割的边界损失。
Med Image Anal. 2021 Jan;67:101851. doi: 10.1016/j.media.2020.101851. Epub 2020 Oct 6.
3
UNet++: A Nested U-Net Architecture for Medical Image Segmentation.U-Net++:一种用于医学图像分割的嵌套U-Net架构。
Deep Learn Med Image Anal Multimodal Learn Clin Decis Support (2018). 2018 Sep;11045:3-11. doi: 10.1007/978-3-030-00889-5_1. Epub 2018 Sep 20.
4
MultiResUNet : Rethinking the U-Net architecture for multimodal biomedical image segmentation.多模态生物医学图像分割的 U-Net 架构再思考:MultiResUNet
Neural Netw. 2020 Jan;121:74-87. doi: 10.1016/j.neunet.2019.08.025. Epub 2019 Sep 4.
5
Attention gated networks: Learning to leverage salient regions in medical images.注意门控网络:学习利用医学图像中的显著区域。
Med Image Anal. 2019 Apr;53:197-207. doi: 10.1016/j.media.2019.01.012. Epub 2019 Feb 5.
6
DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs.DeepLab:基于深度卷积网络、空洞卷积和全连接条件随机场的语义图像分割。
IEEE Trans Pattern Anal Mach Intell. 2018 Apr;40(4):834-848. doi: 10.1109/TPAMI.2017.2699184. Epub 2017 Apr 27.