Pak Sehyun, Son Hye Joo, Kim Dongwoo, Woo Ji Young, Yang Ik, Hwang Hee Sung, Rim Dohyoung, Choi Min Seok, Lee Suk Hyun
Department of Medicine, Hallym University College of Medicine, Chuncheon, Gangwon, Republic of Korea.
Department of Nuclear Medicine, Dankook University Medical Center, Cheonan, Chungnam, Republic of Korea.
Clin Nucl Med. 2025 Jul 1;50(7):596-604. doi: 10.1097/RLU.0000000000005898. Epub 2025 Apr 16.
Convolutional neural networks (CNNs) have been studied for detecting bone metastases on bone scans; however, the application of ConvNeXt and transformer models has not yet been explored. This study aims to evaluate the performance of various deep learning models, including the ConvNeXt and transformer models, in diagnosing metastatic lesions from bone scans.
We retrospectively analyzed bone scans from patients with cancer obtained at 2 institutions: the training and validation sets (n=4626) were from Hospital 1 and the test set (n=1428) was from Hospital 2. The deep learning models evaluated included ResNet18, the Data-Efficient Image Transformer (DeiT), the Vision Transformer (ViT Large 16), the Swin Transformer (Swin Base), and ConvNeXt Large. Gradient-weighted class activation mapping (Grad-CAM) was used for visualization.
Both the validation set and the test set demonstrated that the ConvNeXt large model (0.969 and 0.885, respectively) exhibited the best performance, followed by the Swin Base model (0.965 and 0.840, respectively), both of which significantly outperformed ResNet (0.892 and 0.725, respectively). Subgroup analyses revealed that all the models demonstrated greater diagnostic accuracy for patients with polymetastasis compared with those with oligometastasis. Grad-CAM visualization revealed that the ConvNeXt Large model focused more on identifying local lesions, whereas the Swin Base model focused on global areas such as the axial skeleton and pelvis.
Compared with traditional CNN and transformer models, the ConvNeXt model demonstrated superior diagnostic performance in detecting bone metastases from bone scans, especially in cases of polymetastasis, suggesting its potential in medical image analysis.
卷积神经网络(CNN)已被用于骨扫描中骨转移灶的检测;然而,ConvNeXt和Transformer模型的应用尚未得到探索。本研究旨在评估包括ConvNeXt和Transformer模型在内的各种深度学习模型在诊断骨扫描转移灶中的性能。
我们回顾性分析了在两家机构获取的癌症患者的骨扫描:训练集和验证集(n = 4626)来自医院1,测试集(n = 1428)来自医院2。评估的深度学习模型包括ResNet18、数据高效图像变换器(DeiT)、视觉变换器(ViT Large 16)、Swin变换器(Swin Base)和ConvNeXt Large。采用梯度加权类激活映射(Grad-CAM)进行可视化。
验证集和测试集均显示,ConvNeXt large模型(分别为0.969和0.885)表现最佳,其次是Swin Base模型(分别为0.965和0.840),这两个模型均显著优于ResNet(分别为0.892和0.725)。亚组分析显示,与寡转移患者相比,所有模型对多转移患者的诊断准确性更高。Grad-CAM可视化显示,ConvNeXt Large模型更专注于识别局部病变,而Swin Base模型则专注于轴向骨骼和骨盆等全局区域。
与传统的CNN和Transformer模型相比,ConvNeXt模型在检测骨扫描中的骨转移灶方面表现出卓越的诊断性能,尤其是在多转移病例中,表明其在医学图像分析中的潜力。