Mohammed Foziya Ahmed, Tune Kula Kekeba, Mohammed Juhar Ahmed, Wassu Tizazu Alemu, Muhie Seid
Department of Software Engineering, College of Engineering, Addis Ababa Science and Technology University, Addis Ababa 16417, Ethiopia.
Center of Excellence for HPC and Big Data Analytics, Addis Ababa Science and Technology University, Addis Ababa 16417, Ethiopia.
Diagnostics (Basel). 2024 Oct 14;14(20):2286. doi: 10.3390/diagnostics14202286.
Early diagnosis of cervical cancer at the precancerous stage is critical for effective treatment and improved patient outcomes. This study aims to explore the use of SWIN Transformer and Convolutional Neural Network (CNN) hybrid models combined with transfer learning to classify precancerous colposcopy images. Out of 913 images from 200 cases obtained from the Colposcopy Image Bank of the International Agency for Research on Cancer, 898 met quality standards and were classified as normal, precancerous, or cancerous based on colposcopy and histopathological findings. The cases corresponding to the 360 precancerous images, along with an equal number of normal cases, were divided into a 70/30 train-test split. The SWIN Transformer and CNN hybrid model combines the advantages of local feature extraction by CNNs with the global context modeling by SWIN Transformers, resulting in superior classification performance and a more automated process. The hybrid model approach involves enhancing image quality through preprocessing, extracting local features with CNNs, capturing the global context with the SWIN Transformer, integrating these features for classification, and refining the training process by tuning hyperparameters. The trained model achieved the following classification performances on fivefold cross-validation data: a 94% Area Under the Curve (AUC), an 88% F1 score, and 87% accuracy. On two completely independent test sets, which were never seen by the model during training, the model achieved an 80% AUC, a 75% F1 score, and 75% accuracy on the first test set (precancerous vs. normal) and an 82% AUC, a 78% F1 score, and 75% accuracy on the second test set (cancer vs. normal). : These high-performance metrics demonstrate the models' effectiveness in distinguishing precancerous from normal colposcopy images, even with modest datasets, limited data augmentation, and the smaller effect size of precancerous images compared to malignant lesions. The findings suggest that these techniques can significantly aid in the early detection of cervical cancer at the precancerous stage.
宫颈癌癌前阶段的早期诊断对于有效治疗和改善患者预后至关重要。本研究旨在探索使用SWIN Transformer和卷积神经网络(CNN)混合模型结合迁移学习对癌前阴道镜图像进行分类。从国际癌症研究机构的阴道镜图像库中获取的200例患者的913张图像中,898张符合质量标准,并根据阴道镜检查和组织病理学结果分为正常、癌前或癌性。与360张癌前图像对应的病例,以及相同数量的正常病例,被分为70/30的训练-测试分割。SWIN Transformer和CNN混合模型结合了CNN局部特征提取的优势和SWIN Transformer的全局上下文建模,从而产生了卓越的分类性能和更自动化的过程。混合模型方法包括通过预处理提高图像质量,用CNN提取局部特征,用SWIN Transformer捕捉全局上下文,整合这些特征进行分类,并通过调整超参数优化训练过程。训练后的模型在五折交叉验证数据上实现了以下分类性能:曲线下面积(AUC)为94%,F1分数为88%,准确率为87%。在两个完全独立的测试集上,即模型在训练期间从未见过的数据,模型在第一个测试集(癌前与正常)上实现了80%的AUC、75%的F1分数和75%的准确率,在第二个测试集(癌与正常)上实现了82%的AUC、78%的F1分数和75%的准确率。这些高性能指标表明,即使数据集规模不大、数据增强有限且癌前图像与恶性病变相比效应量较小,这些模型在区分癌前阴道镜图像与正常图像方面也很有效。研究结果表明,这些技术可以显著有助于宫颈癌癌前阶段的早期检测。