Zhu Yixin, Wu Ji, Long Qiongxian, Li Yan, Luo Hao, Pang Lu, Zhu Lin, Luo Hui
Department of Ultrasound, Peking University Shenzhen Hospital, Shenzhen, Guangdong, China.
Department of Urology, The Affiliated Nanchong Central Hospital of North Sichuan Medical College, Nanchong, Sichuan, China.
Front Physiol. 2025 Mar 7;16:1558997. doi: 10.3389/fphys.2025.1558997. eCollection 2025.
This study aimed to develop and validate a multimodal deep learning model that utilizes preoperative grayscale and contrast-enhanced ultrasound (CEUS) video data for noninvasive WHO/ISUP nuclear grading of renal cell carcinoma (RCC).
In this dual-center retrospective study, CEUS videos from 100 patients with RCC collected between June 2012 and June 2021 were analyzed. A total of 6,293 ultrasound images were categorized into low-grade (G1-G2) and high-grade (G3-G4) groups. A novel model, the Multimodal Ultrasound Fusion Network (MUF-Net), integrated B-mode and CEUS modalities to extract and fuse image features using a weighted sum of predicted weights. Model performance was assessed using five-fold cross-validation and compared to single-modality models. Grad-CAM visualization highlighted key regions influencing the model's predictions.
MUF-Net achieved an accuracy of 85.9%, outperforming B-mode (80.8%) and CEUS-mode (81.8%, < 0.05) models. Sensitivities were 85.1%, 80.2%, and 77.8%, while specificities were 86.0%, 82.5%, and 82.7%, respectively. The AUC of MUF-Net (0.909, 95% CI: 0.829-0.990) was superior to B-mode (0.838, 95% CI: 0.689-0.988) and CEUS-mode (0.845, 95% CI: 0.745-0.944). Grad-CAM analysis revealed distinct and complementary salient regions across modalities.
MUF-Net provides accurate and interpretable RCC nuclear grading, surpassing unimodal approaches, with Grad-CAM offering intuitive insights into the model's predictions.
本研究旨在开发并验证一种多模态深度学习模型,该模型利用术前灰度和对比增强超声(CEUS)视频数据对肾细胞癌(RCC)进行无创的世界卫生组织/国际泌尿病理学会(WHO/ISUP)核分级。
在这项双中心回顾性研究中,分析了2012年6月至2021年6月期间收集的100例RCC患者的CEUS视频。总共6293张超声图像被分为低级别(G1-G2)和高级别(G3-G4)组。一种新型模型,即多模态超声融合网络(MUF-Net),整合了B模式和CEUS模态,以使用预测权重的加权和来提取和融合图像特征。使用五折交叉验证评估模型性能,并与单模态模型进行比较。Grad-CAM可视化突出了影响模型预测的关键区域。
MUF-Net的准确率达到85.9%,优于B模式(80.8%)和CEUS模式(81.8%,P<0.05)模型。敏感性分别为85.1%、80.2%和77.8%,特异性分别为86.0%、82.5%和82.7%。MUF-Net的曲线下面积(AUC)(0.909,95%置信区间:0.829-0.990)优于B模式(0.838,95%置信区间:0.689-0.988)和CEUS模式(0.845,95%置信区间:0.745-0.944)。Grad-CAM分析揭示了不同模态之间独特且互补的显著区域。
MUF-Net提供了准确且可解释的RCC核分级,优于单模态方法,Grad-CAM为模型预测提供了直观的见解。