Abdikenov Beibit, Zhaksylyk Tomiris, Imasheva Aruzhan, Orazayev Yerzhan, Karibekov Temirlan
Science and Innovation Center "Artificial Intelligence", Astana IT University, Astana 010000, Kazakhstan.
J Imaging. 2025 Jul 22;11(8):247. doi: 10.3390/jimaging11080247.
Mammography is the main method for early detection of breast cancer, which is still a major global health concern. However, inter-reader variability and the inherent difficulty of interpreting subtle radiographic features frequently limit the accuracy of diagnosis. A thorough assessment of deep convolutional neural networks (CNNs) for automated mammogram classification is presented in this work, along with the introduction of two innovative multi-view integration techniques: Dual-Branch Ensemble (DBE) and Merged Dual-View (MDV). By setting aside two datasets for out-of-sample testing, we evaluate the generalizability of the model using six different mammography datasets that represent various populations and imaging systems. We compare a number of cutting-edge architectures on both individual and combined datasets, including ResNet, DenseNet, EfficientNet, MobileNet, Vision Transformers, and VGG19. Both MDV and DBE strategies improve classification performance, according to experimental results. VGG19 and DenseNet both obtained high ROC AUC scores of 0.9051 and 0.7960 under the MDV approach. DenseNet demonstrated strong performance in the DBE setting, achieving a ROC AUC of 0.8033, while ResNet50 recorded a ROC AUC of 0.8042. These enhancements demonstrate how beneficial multi-view fusion is for boosting model robustness. The impact of domain shift is further highlighted by generalization tests, which emphasize the need for diverse datasets in training. These results offer practical advice for improving CNN architectures and integration tactics, which will aid in the creation of trustworthy, broadly applicable AI-assisted breast cancer screening tools.
乳腺钼靶摄影是早期检测乳腺癌的主要方法,而乳腺癌仍是全球主要的健康问题。然而,不同阅片者之间的差异以及解读细微影像学特征的固有难度常常限制了诊断的准确性。本文对用于自动乳腺钼靶摄影分类的深度卷积神经网络(CNN)进行了全面评估,并介绍了两种创新的多视图整合技术:双分支集成(DBE)和合并双视图(MDV)。通过留出两个数据集用于样本外测试,我们使用六个代表不同人群和成像系统的不同乳腺钼靶摄影数据集评估了模型的泛化能力。我们在单个和组合数据集上比较了许多前沿架构,包括ResNet、DenseNet、EfficientNet、MobileNet、视觉Transformer和VGG19。实验结果表明,MDV和DBE策略均提高了分类性能。在MDV方法下,VGG19和DenseNet的ROC AUC得分分别高达0.9051和0.7960。DenseNet在DBE设置中表现出色,ROC AUC为0.8033,而ResNet50的ROC AUC为0.8042。这些改进表明了多视图融合对提高模型鲁棒性的益处。泛化测试进一步凸显了域转移的影响,强调了训练中使用多样化数据集的必要性。这些结果为改进CNN架构和整合策略提供了实用建议,这将有助于创建可靠、广泛适用的人工智能辅助乳腺癌筛查工具。