Banerjee Soumyanil, Summerfield Nicholas, Dong Ming, Glide-Hurst Carri
IEEE Trans Biomed Eng. 2025 May 5;PP. doi: 10.1109/TBME.2025.3566995.
U-shaped networks and its variants have demonstrated exceptional results for medical image segmentation. In this paper, we propose a novel dual self-distillation (DSD) framework in U-shaped networks for volumetric medical image segmentation. DSD distills knowledge from the ground-truth segmentation labels to the decoder layers. Additionally, DSD also distills knowledge from the deepest decoder and encoder layer to the shallower decoder and encoder layers respectively of a single U-shaped network. DSD is a general training strategy that could be attached to the backbone architecture of any U-shaped network to further improve its segmentation performance. We attached DSD on several state-of-the-art U-shaped backbones, and extensive experiments on various public 3D medical image segmentation datasets (cardiac substructure, brain tumor and Hippocampus) demonstrated significant improvement over the same backbones without DSD. On average, after attaching DSD to the U-shaped backbones, we observed an increase of 2.82%, 4.53% and 1.3% in Dice similarity score, a decrease of 7.15 mm, 6.48 mm and 0.76 mm in the Hausdorff distance, for cardiac substructure, brain tumor and Hippocampus segmentation, respectively. These improvements were achieved with negligible increase in the number of trainable parameters and training time. Our proposed DSD framework also led to significant qualitative improvements for cardiac substructure, brain tumor and Hippocampus segmentation over the U-shaped backbones. The source code is publicly available at https://github.com/soumbane/DualSelfDistillation.
U型网络及其变体在医学图像分割方面已展现出卓越的成果。在本文中,我们提出了一种用于容积医学图像分割的U型网络中的新型双自蒸馏(DSD)框架。DSD将真实分割标签中的知识蒸馏到解码器层。此外,DSD还分别将单个U型网络中最深的解码器和编码器层的知识蒸馏到较浅的解码器和编码器层。DSD是一种通用的训练策略,可以附加到任何U型网络的骨干架构上,以进一步提高其分割性能。我们将DSD附加到几个先进的U型骨干网络上,并且在各种公共3D医学图像分割数据集(心脏亚结构、脑肿瘤和海马体)上进行的广泛实验表明,与没有DSD的相同骨干网络相比有显著改进。平均而言,在将DSD附加到U型骨干网络后,我们观察到在心脏亚结构、脑肿瘤和海马体分割中,Dice相似性分数分别提高了2.82%、4.53%和1.3%,豪斯多夫距离分别减少了7.15毫米、6.48毫米和0.76毫米。这些改进是在可训练参数数量和训练时间增加可忽略不计 的情况下实现的。我们提出的DSD框架在心脏亚结构、脑肿瘤和海马体分割方面相对于U型骨干网络也带来了显著的定性改进。源代码可在https://github.com/soumbane/DualSelfDistillation上公开获取。