Faizi Mohammad Khalid, Qiang Yan, Wei Yangyang, Qiao Ying, Zhao Juanjuan, Aftab Rukhma, Urrehman Zia
College of Computer Science and Technology (College of Data Science), Taiyuan University of Technology, Taiyuan, 030024, Shanxi, China.
School of Software, North University of China, Taiyuan, Shanxi, China.
BMC Cancer. 2025 Jul 1;25(1):1056. doi: 10.1186/s12885-025-14320-8.
Lung cancer remains a leading cause of cancer-related deaths worldwide, with accurate classification of lung nodules being critical for early diagnosis. Traditional radiological methods often struggle with high false-positive rates, underscoring the need for advanced diagnostic tools. In this work, we introduce DCSwinB, a novel deep learning-based lung nodule classifier designed to improve the accuracy and efficiency of benign and malignant nodule classification in CT images. Built on the Swin-Tiny Vision Transformer (ViT), DCSwinB incorporates several key innovations: a dual-branch architecture that combines CNNs for local feature extraction and Swin Transformer for global feature extraction, and a Conv-MLP module that enhances connections between adjacent windows to capture long-range dependencies in 3D images. Pretrained on the LUNA16 and LUNA16-K datasets, which consist of annotated CT scans from thousands of patients, DCSwinB was evaluated using ten-fold cross-validation. The model demonstrated superior performance, achieving 90.96% accuracy, 90.56% recall, 89.65% specificity, and an AUC of 0.94, outperforming existing models such as ResNet50 and Swin-T. These results highlight the effectiveness of DCSwinB in enhancing feature representation while optimizing computational efficiency. By improving the accuracy and reliability of lung nodule classification, DCSwinB has the potential to assist radiologists in reducing diagnostic errors, enabling earlier intervention and improved patient outcomes.
肺癌仍然是全球癌症相关死亡的主要原因,准确分类肺结节对于早期诊断至关重要。传统的放射学方法常常面临高假阳性率的问题,这凸显了对先进诊断工具的需求。在这项工作中,我们引入了DCSwinB,这是一种基于深度学习的新型肺结节分类器,旨在提高CT图像中良性和恶性结节分类的准确性和效率。DCSwinB基于Swin-Tiny视觉Transformer(ViT)构建,融合了多项关键创新:一种双分支架构,结合了用于局部特征提取的卷积神经网络(CNN)和用于全局特征提取的Swin Transformer;以及一个Conv-MLP模块,该模块增强相邻窗口之间的连接以捕捉三维图像中的长程依赖性。DCSwinB在由数千名患者的标注CT扫描组成的LUNA16和LUNA16-K数据集上进行预训练,并使用十折交叉验证进行评估。该模型表现出卓越的性能,准确率达到90.96%,召回率达到90.56%,特异性达到89.65%,曲线下面积(AUC)为0.94,优于诸如ResNet50和Swin-T等现有模型。这些结果突出了DCSwinB在增强特征表示同时优化计算效率方面的有效性。通过提高肺结节分类的准确性和可靠性,DCSwinB有潜力帮助放射科医生减少诊断错误,实现更早的干预并改善患者预后。