Chari Shruthi, Acharya Prasant, Gruen Daniel M, Zhang Olivia, Eyigoz Elif K, Ghalwash Mohamed, Seneviratne Oshani, Saiz Fernando Suarez, Meyer Pablo, Chakraborty Prithwish, McGuinness Deborah L
Rensselaer Polytechnic Institute, 110 8th St, Troy, 12180, NY, USA.
Rensselaer Polytechnic Institute, 110 8th St, Troy, 12180, NY, USA.
Artif Intell Med. 2023 Mar;137:102498. doi: 10.1016/j.artmed.2023.102498. Epub 2023 Feb 2.
Medical experts may use Artificial Intelligence (AI) systems with greater trust if these are supported by 'contextual explanations' that let the practitioner connect system inferences to their context of use. However, their importance in improving model usage and understanding has not been extensively studied. Hence, we consider a comorbidity risk prediction scenario and focus on contexts regarding the patients' clinical state, AI predictions about their risk of complications, and algorithmic explanations supporting the predictions. We explore how relevant information for such dimensions can be extracted from Medical guidelines to answer typical questions from clinical practitioners. We identify this as a question answering (QA) task and employ several state-of-the-art Large Language Models (LLM) to present contexts around risk prediction model inferences and evaluate their acceptability. Finally, we study the benefits of contextual explanations by building an end-to-end AI pipeline including data cohorting, AI risk modeling, post-hoc model explanations, and prototyped a visual dashboard to present the combined insights from different context dimensions and data sources, while predicting and identifying the drivers of risk of Chronic Kidney Disease (CKD) - a common type-2 diabetes (T2DM) comorbidity. All of these steps were performed in deep engagement with medical experts, including a final evaluation of the dashboard results by an expert medical panel. We show that LLMs, in particular BERT and SciBERT, can be readily deployed to extract some relevant explanations to support clinical usage. To understand the value-add of the contextual explanations, the expert panel evaluated these regarding actionable insights in the relevant clinical setting. Overall, our paper is one of the first end-to-end analyses identifying the feasibility and benefits of contextual explanations in a real-world clinical use case. Our findings can help improve clinicians' usage of AI models.
如果人工智能(AI)系统能得到“上下文解释”的支持,使从业者能够将系统推理与使用背景联系起来,医学专家可能会更信任地使用这些系统。然而,它们在改善模型使用和理解方面的重要性尚未得到广泛研究。因此,我们考虑一种合并症风险预测场景,并关注患者临床状态、AI对其并发症风险的预测以及支持这些预测的算法解释等方面。我们探索如何从医学指南中提取这些维度的相关信息,以回答临床医生的典型问题。我们将此识别为一个问答(QA)任务,并使用几个最先进的大语言模型(LLM)来呈现风险预测模型推理周围的上下文,并评估其可接受性。最后,我们通过构建一个端到端的AI管道来研究上下文解释的好处,该管道包括数据分组、AI风险建模、事后模型解释,并制作了一个可视化仪表板原型,以呈现来自不同上下文维度和数据源的综合见解,同时预测和识别慢性肾脏病(CKD)——一种常见的2型糖尿病(T2DM)合并症的风险驱动因素。所有这些步骤都是在与医学专家深入合作的情况下进行的,包括由专家医学小组对仪表板结果进行最终评估。我们表明,大语言模型,特别是BERT和SciBERT,可以很容易地部署来提取一些相关解释以支持临床使用。为了理解上下文解释的附加值,专家小组在相关临床环境中就可操作的见解对这些解释进行了评估。总体而言,我们的论文是最早对现实世界临床用例中上下文解释的可行性和好处进行的端到端分析之一。我们的发现有助于改善临床医生对AI模型的使用情况。