Charlton Colleen E, Poon Michael T C, Brennan Paul M, Fleuriot Jacques D
Artificial Intelligence and its Applications Institute, School of Informatics, University of Edinburgh, 10 Crichton Street, Edinburgh EH8 9AB, UK.
Cancer Research UK Brain Tumour Centre of Excellence, CRUK Edinburgh Centre, University of Edinburgh, Edinburgh, UK; Department of Clinical Neuroscience, Royal Infirmary of Edinburgh, 51 Little France Crescent EH16 4SA, UK.; Translational Neurosurgery, Centre for Clinical Brain Sciences, University of Edinburgh, Edinburgh, UK; Centre for Medical Informatics, Usher Institute, University of Edinburgh, Edinburgh, UK.
Comput Methods Programs Biomed. 2023 May;233:107482. doi: 10.1016/j.cmpb.2023.107482. Epub 2023 Mar 13.
Prediction of survival in patients diagnosed with a brain tumour is challenging because of heterogeneous tumour behaviours and treatment response. Advances in machine learning have led to the development of clinical prognostic models, but due to the lack of model interpretability, integration into clinical practice is almost non-existent. In this retrospective study, we compare five classification models with varying degrees of interpretability for the prediction of brain tumour survival greater than one year following diagnosis.
1028 patients aged ≥16 years with a brain tumour diagnosis between April 2012 and April 2020 were included in our study. Three intrinsically interpretable 'glass box' classifiers (Bayesian Rule Lists [BRL], Explainable Boosting Machine [EBM], and Logistic Regression [LR]), and two 'black box' classifiers (Random Forest [RF] and Support Vector Machine [SVM]) were trained on electronic patients records for the prediction of one-year survival. All models were evaluated using balanced accuracy (BAC), F1-score, sensitivity, specificity, and receiver operating characteristics. Black box model interpretability and misclassified predictions were quantified using SHapley Additive exPlanations (SHAP) values and model feature importance was evaluated by clinical experts.
The RF model achieved the highest BAC of 78.9%, closely followed by SVM (77.7%), LR (77.5%) and EBM (77.1%). Across all models, age, diagnosis (tumour type), functional features, and first treatment were top contributors to the prediction of one year survival. We used EBM and SHAP to explain model misclassifications and investigated the role of feature interactions in prognosis.
Interpretable models are a natural choice for the domain of predictive medicine. Intrinsically interpretable models, such as EBMs, may provide an advantage over traditional clinical assessment of brain tumour prognosis by weighting potential risk factors and their interactions that may be unknown to clinicians. An agreement between model predictions and clinical knowledge is essential for establishing trust in the models decision making process, as well as trust that the model will make accurate predictions when applied to new data.
由于脑肿瘤行为和治疗反应的异质性,预测脑肿瘤患者的生存率具有挑战性。机器学习的进展促使了临床预后模型的发展,但由于缺乏模型可解释性,其几乎无法融入临床实践。在这项回顾性研究中,我们比较了五种具有不同程度可解释性的分类模型,以预测脑肿瘤诊断后一年以上的生存率。
我们的研究纳入了2012年4月至2020年4月期间诊断为脑肿瘤的1028例年龄≥16岁的患者。在电子病历上训练了三种内在可解释的“玻璃盒”分类器(贝叶斯规则列表[BRL]、可解释增强机器[EBM]和逻辑回归[LR])以及两种“黑盒”分类器(随机森林[RF]和支持向量机[SVM]),以预测一年生存率。所有模型均使用平衡准确率(BAC)、F1分数、灵敏度、特异性和受试者工作特征进行评估。使用SHapley加性解释(SHAP)值对黑盒模型的可解释性和错误分类预测进行量化,并由临床专家评估模型特征重要性。
RF模型的BAC最高,为78.9%,紧随其后的是SVM(77.7%)、LR(77.5%)和EBM(77.1%)。在所有模型中,年龄、诊断(肿瘤类型)、功能特征和首次治疗是预测一年生存率的主要因素。我们使用EBM和SHAP来解释模型的错误分类,并研究特征相互作用在预后中的作用。
可解释模型是预测医学领域的自然选择。内在可解释的模型,如EBM,通过权衡临床医生可能未知的潜在风险因素及其相互作用,可能比传统的脑肿瘤预后临床评估具有优势。模型预测与临床知识之间的一致性对于建立对模型决策过程的信任至关重要,同时也有助于信任模型在应用于新数据时能够做出准确的预测。