Zhu Chengzhang, Chai Xian, Xiao Yalong, Liu Xu, Zhang Renmao, Yang Zhangzheng, Wang Zhiyuan
School of Humanities, Central South University, Changsha 410012, China.
School of Computer Science and Engineering, Central South University, Changsha 410083, China.
Diagnostics (Basel). 2024 Jan 26;14(3):269. doi: 10.3390/diagnostics14030269.
Breast cancer is one of the most common cancers in the world, especially among women. Breast tumor segmentation is a key step in the identification and localization of the breast tumor region, which has important clinical significance. Inspired by the swin-transformer model with powerful global modeling ability, we propose a semantic segmentation framework named Swin-Net for breast ultrasound images, which combines Transformer and Convolutional Neural Networks (CNNs) to effectively improve the accuracy of breast ultrasound segmentation. Firstly, our model utilizes a swin-transformer encoder with stronger learning ability, which can extract features of images more precisely. In addition, two new modules are introduced in our method, including the feature refinement and enhancement module (RLM) and the hierarchical multi-scale feature fusion module (HFM), given that the influence of ultrasonic image acquisition methods and the characteristics of tumor lesions is difficult to capture. Among them, the RLM module is used to further refine and enhance the feature map learned by the transformer encoder. The HFM module is used to process multi-scale high-level semantic features and low-level details, so as to achieve effective cross-layer feature fusion, suppress noise, and improve model segmentation performance. Experimental results show that Swin-Net performs significantly better than the most advanced methods on the two public benchmark datasets. In particular, it achieves an absolute improvement of 1.4-1.8% on Dice. Additionally, we provide a new dataset of breast ultrasound images on which we test the effect of our model, further demonstrating the validity of our method. In summary, the proposed Swin-Net framework makes significant advancements in breast ultrasound image segmentation, providing valuable exploration for research and applications in this domain.
乳腺癌是世界上最常见的癌症之一,在女性中尤为如此。乳腺肿瘤分割是识别和定位乳腺肿瘤区域的关键步骤,具有重要的临床意义。受具有强大全局建模能力的Swin-Transformer模型启发,我们提出了一种名为Swin-Net的乳腺超声图像语义分割框架,该框架结合了Transformer和卷积神经网络(CNN),以有效提高乳腺超声分割的准确性。首先,我们的模型利用了具有更强学习能力的Swin-Transformer编码器,它可以更精确地提取图像特征。此外,由于超声图像采集方法的影响和肿瘤病变特征难以捕捉,我们的方法中引入了两个新模块,包括特征细化与增强模块(RLM)和分层多尺度特征融合模块(HFM)。其中,RLM模块用于进一步细化和增强由Transformer编码器学习到的特征图。HFM模块用于处理多尺度的高级语义特征和低级细节,从而实现有效的跨层特征融合,抑制噪声,并提高模型分割性能。实验结果表明,在两个公开的基准数据集上,Swin-Net的表现明显优于最先进的方法。特别是,它在Dice上实现了1.4-1.8%的绝对提升。此外,我们提供了一个新的乳腺超声图像数据集,在该数据集上测试了我们模型的效果,进一步证明了我们方法的有效性。总之,所提出的Swin-Net框架在乳腺超声图像分割方面取得了显著进展,为该领域的研究和应用提供了有价值的探索。