Jiang Chong, Jiang Zekun, Zhang Zitong, Huang Hexiao, Zhou Hang, Jiang Qiuhui, Teng Yue, Li Hai, Xu Bing, Li Xin, Xu Jingyan, Ding Chongyang, Li Kang, Tian Rong
Department of Nuclear Medicine, West China Hospital, Sichuan University, No.37, Guoxue Alley, Chengdu City, Sichuan Province, 610041, China.
West China Biomedical Big Data Center, West China Hospital, Sichuan University, No.37, Guoxue Alley, Chengdu City, Sichuan Province, 610041, China.
Eur J Nucl Med Mol Imaging. 2025 Jun;52(7):2384-2396. doi: 10.1007/s00259-025-07090-9. Epub 2025 Jan 30.
Pathological grade is a critical determinant of clinical outcomes and decision-making of follicular lymphoma (FL). This study aimed to develop a deep learning model as a digital biopsy for the non-invasive identification of FL grade.
This study retrospectively included 513 FL patients from five independent hospital centers, randomly divided into training, internal validation, and external validation cohorts. A multimodal fusion Transformer model was developed integrating 3D PET tumor images with tabular data to predict FL grade. Additionally, the model is equipped with explainable modules, including Gradient-weighted Class Activation Mapping (Grad-CAM) for PET images, SHapley Additive exPlanations analysis for tabular data, and the calculation of predictive contribution ratios for both modalities, to enhance clinical interpretability and reliability. The predictive performance was evaluated using the area under the receiver operating characteristic curve (AUC) and accuracy, and its prognostic value was also assessed.
The Transformer model demonstrated high accuracy in grading FL, with AUCs of 0.964-0.985 and accuracies of 90.2-96.7% in the training cohort, and similar performance in the validation cohorts (AUCs: 0.936-0.971, accuracies: 86.4-97.0%). Ablation studies confirmed that the fusion model outperformed single-modality models (AUCs: 0.974 - 0.956, accuracies: 89.8%-85.8%). Interpretability analysis revealed that PET images contributed 81-89% of the predictive value. Grad-CAM highlighted the tumor and peri-tumor regions. The model also effectively stratified patients by survival risk (P < 0.05), highlighting its prognostic value.
Our study developed an explainable multimodal fusion Transformer model for accurate grading and prognosis of FL, with the potential to aid clinical decision-making.
病理分级是滤泡性淋巴瘤(FL)临床结局和决策的关键决定因素。本研究旨在开发一种深度学习模型作为数字活检工具,用于无创识别FL分级。
本研究回顾性纳入了来自五个独立医院中心的513例FL患者,随机分为训练组、内部验证组和外部验证组。开发了一种多模态融合Transformer模型,将3D PET肿瘤图像与表格数据相结合以预测FL分级。此外,该模型配备了可解释模块,包括用于PET图像的梯度加权类激活映射(Grad-CAM)、用于表格数据的SHapley加性解释分析以及两种模态的预测贡献比计算,以提高临床可解释性和可靠性。使用受试者操作特征曲线下面积(AUC)和准确性评估预测性能,并评估其预后价值。
Transformer模型在FL分级方面表现出高准确性,训练组的AUC为0.964 - 0.985,准确性为90.2 - 96.7%,验证组的表现相似(AUC:0.936 - 0.971,准确性:86.4 - 97.0%)。消融研究证实融合模型优于单模态模型(AUC:0.974 - 0.956,准确性:89.8% - 85.8%)。可解释性分析表明PET图像贡献了81 - 89%的预测价值。Grad-CAM突出了肿瘤和肿瘤周围区域。该模型还能有效地按生存风险对患者进行分层(P < 0.05),突出了其预后价值。
我们的研究开发了一种可解释的多模态融合Transformer模型,用于准确分级和预测FL的预后,并有可能辅助临床决策。