Shen Quanyou, Zheng Bowen, Li Wenhao, Shi Xiaoran, Luo Kun, Yao Yuqian, Li Xinyan, Lv Shidong, Tao Jie, Wei Qiang
School of Automation, Guangdong University of Technology, Guangzhou, 510006, China; Guangdong Provincial Key Laboratory of Intelligent Decision and Cooperative Control, Guangzhou, 510006, China; Guangdong-Hong Kong Joint Laboratory for Intelligent Decision and Cooperative Control, Guangzhou, 510006, China.
Department of Urology, Nanfang Hospital, Southern Medical University, Guangzhou, 510515, China.
Neural Netw. 2025 Jan;181:106782. doi: 10.1016/j.neunet.2024.106782. Epub 2024 Oct 5.
Magnetic resonance imaging (MRI) plays a pivotal role in diagnosing and staging prostate cancer. Precise delineation of the peripheral zone (PZ) and transition zone (TZ) within prostate MRI is essential for accurate diagnosis and subsequent artificial intelligence-driven analysis. However, existing segmentation methods are limited by ambiguous boundaries, shape variations and texture complexities between PZ and TZ. Moreover, they suffer from inadequate modeling capabilities and limited receptive fields. To address these challenges, we propose a Enhanced MixFormer, which integrates window-based multi-head self-attention (W-MSA) and depth-wise convolution with parallel design and cross-branch bidirectional interaction. We further introduce MixUNETR, which use multiple Enhanced MixFormers as encoder to extract features from both PZ and TZ in prostate MRI. This augmentation effectively enlarges the receptive field and enhances the modeling capability of W-MSA, ultimately improving the extraction of both global and local feature information from PZ and TZ, thereby addressing mis-segmentation and challenges in delineating boundaries between them. Extensive experiments were conducted, comparing MixUNETR with several state-of-the-art methods on the Prostate158, ProstateX public datasets and private dataset. The results consistently demonstrate the accuracy and robustness of MixUNETR in MRI prostate segmentation. Our code of methods is available at https://github.com/skyous779/MixUNETR.git.
磁共振成像(MRI)在前列腺癌的诊断和分期中起着关键作用。在前列腺MRI中精确描绘外周带(PZ)和移行带(TZ)对于准确诊断以及后续的人工智能驱动分析至关重要。然而,现有的分割方法受到PZ和TZ之间边界模糊、形状变化和纹理复杂性的限制。此外,它们还存在建模能力不足和感受野有限的问题。为了应对这些挑战,我们提出了一种增强型混合变换器(Enhanced MixFormer),它将基于窗口的多头自注意力(W-MSA)和深度卷积进行并行设计并交叉分支双向交互。我们进一步引入了混合U-Net变换器(MixUNETR),它使用多个增强型混合变换器作为编码器,从前列腺MRI的PZ和TZ中提取特征。这种增强有效地扩大了感受野并增强了W-MSA的建模能力,最终改善了从PZ和TZ中提取全局和局部特征信息的效果,从而解决了分割错误以及描绘它们之间边界的挑战。我们进行了广泛的实验,在Prostate158、ProstateX公共数据集和私有数据集上,将MixUNETR与几种最先进的方法进行了比较。结果一致证明了MixUNETR在MRI前列腺分割中的准确性和鲁棒性。我们的方法代码可在https://github.com/skyous779/MixUNETR.git获取。