Zhang Lin, Yin Xiaochun, Liu Xuqi, Liu Zengguang
College of Computer Science, Weifang University of Science and Technology, Weifang, 262700, China.
School of Information Engineering, Shandong Vocational College of Science and Technology, Weifang, 261053, China.
Sci Rep. 2025 Apr 25;15(1):14565. doi: 10.1038/s41598-025-97779-6.
Medical image segmentation plays a crucial role in assisting clinical diagnosis, yet existing models often struggle with handling diverse and complex medical data, particularly when dealing with multi-scale organ and tissue structures. This paper proposes a novel medical image segmentation model, FE-SwinUper, designed to address these challenges by integrating the strengths of the Swin Transformer and UPerNet architectures. The objective is to enhance multi-scale feature extraction and improve the fusion of hierarchical organ and tissue representations through a feature enhancement Swin Transformer (FE-ST) backbone and an adaptive feature fusion (AFF) module. The FE-ST backbone utilizes self-attention mechanisms to efficiently extract rich spatial and contextual features across different scales, while the AFF module adapts to multi-scale feature fusion, mitigating the loss of contextual information. We evaluate the model on two publicly available medical image segmentation datasets: Synapse multi-organ segmentation dataset and the ACDC cardiac segmentation dataset. Our results show that FE-SwinUper outperforms existing state-of-the-art models in terms of Dice coefficient, pixel accuracy, and Hausdorff distance. The model achieves a Dice score of 91.58% on the Synapse dataset and 90.15% on the ACDC dataset. These results demonstrate the robustness and efficiency of the proposed model, indicating its potential for real-world clinical applications.
医学图像分割在辅助临床诊断中起着至关重要的作用,但现有模型在处理多样且复杂的医学数据时往往面临困难,尤其是在处理多尺度器官和组织结构时。本文提出了一种新颖的医学图像分割模型FE-SwinUper,旨在通过整合Swin Transformer和UPerNet架构的优势来应对这些挑战。目标是通过特征增强Swin Transformer(FE-ST)主干和自适应特征融合(AFF)模块来增强多尺度特征提取,并改善分层器官和组织表示的融合。FE-ST主干利用自注意力机制在不同尺度上高效提取丰富的空间和上下文特征,而AFF模块则适应多尺度特征融合,减轻上下文信息的损失。我们在两个公开可用的医学图像分割数据集上评估了该模型:Synapse多器官分割数据集和ACDC心脏分割数据集。我们的结果表明,FE-SwinUper在Dice系数、像素准确率和豪斯多夫距离方面优于现有的最先进模型。该模型在Synapse数据集上的Dice分数为91.58%,在ACDC数据集上为90.15%。这些结果证明了所提出模型的稳健性和效率,表明其在实际临床应用中的潜力。