Byeon Haewon, Alsaadi Mahmood, Vijay Richa, Assudani Purshottam J, Kumar Dutta Ashit, Bansal Monika, Singh Pavitar Parkash, Soni Mukesh, Bhatt Mohammed Wasim
Convergence Department, Korea University of Technology and Education, Cheonan, Republic of Korea.
Department of Computer Sciences, College of Sciences, University of Al Maarif, Al Anbar, Iraq.
Front Oncol. 2025 Jul 14;15:1626785. doi: 10.3389/fonc.2025.1626785. eCollection 2025.
Breast cancer remains the most prevalent cancer among women, where accurate and interpretable analysis of pathology images is vital for early diagnosis and personalized treatment planning. However, conventional single-network models fall short in balancing both performance and explainability-Convolutional Neural Networks (CNNs) lack the capacity to capture global contextual information, while Transformers are limited in modeling fine-grained local details. To overcome these challenges and contribute to the advancement of Explainable AI (XAI) in precision cancer diagnosis, this paper proposes MVT-OFML (Multi-View Transformer Online Fusion Mutual Learning), a novel and interpretable classification framework for breast cancer pathology images. MVT-OFML combines ResNet-50 for extracting detailed local features and a multi-view Transformer encoding module for capturing comprehensive global context across multiple perspectives. A key innovation is the Online Fusion Mutual Learning (OFML) mechanism, which enables bidirectional knowledge sharing between the CNN and Transformer branches by aligning both intermediate feature representations and prediction logits. This mutual learning framework enhances performance while also producing interpretable attention maps and feature-level visualizations that reveal the decision-making process of the model-promoting transparency, trust, and clinical usability. Extensive experiments on the BreakHis and BACH datasets demonstrate that MVT-OFML significantly outperforms the strongest baseline models, achieving accuracy improvements of 0.90% and 2.26%, and F-score gains of 4.75% and 3.21%, respectively. By integrating complementary modeling paradigms with explainable learning strategies, MVT-OFML offers a promising AI solution for precise and interpretable breast cancer diagnosis and prognosis, supporting informed decision-making in clinical settings.
乳腺癌仍然是女性中最常见的癌症,对病理图像进行准确且可解释的分析对于早期诊断和个性化治疗规划至关重要。然而,传统的单网络模型在平衡性能和可解释性方面存在不足——卷积神经网络(CNN)缺乏捕捉全局上下文信息的能力,而Transformer在对细粒度局部细节进行建模时存在局限性。为了克服这些挑战并推动可解释人工智能(XAI)在精准癌症诊断中的发展,本文提出了MVT - OFML(多视图Transformer在线融合相互学习),这是一种用于乳腺癌病理图像的新颖且可解释的分类框架。MVT - OFML结合了用于提取详细局部特征的ResNet - 50和用于从多个视角捕捉全面全局上下文的多视图Transformer编码模块。一个关键创新是在线融合相互学习(OFML)机制,它通过对齐中间特征表示和预测逻辑,实现了CNN和Transformer分支之间的双向知识共享。这种相互学习框架提高了性能,同时还生成了可解释的注意力图和特征级可视化,揭示了模型的决策过程——促进了透明度、可信度和临床可用性。在BreakHis和BACH数据集上进行的大量实验表明,MVT - OFML显著优于最强的基线模型,准确率分别提高了0.90%和2.26%,F分数分别提高了4.75%和3.21%。通过将互补的建模范式与可解释的学习策略相结合,MVT - OFML为精确且可解释的乳腺癌诊断和预后提供了一个有前景的人工智能解决方案,支持临床环境中的明智决策。