Department of Pediatrics, Division of Neurology, Children's Healthcare of Atlanta, Emory University School of Medicine, Atlanta, GA, USA.
Georgia Institute of Technology, Atlanta, GA, USA.
Ann Clin Transl Neurol. 2024 May;11(5):1224-1235. doi: 10.1002/acn3.52036. Epub 2024 Apr 5.
OBJECTIVE: Artificial intelligence (AI)-based decision support systems (DSS) are utilized in medicine but underlying decision-making processes are usually unknown. Explainable AI (xAI) techniques provide insight into DSS, but little is known on how to design xAI for clinicians. Here we investigate the impact of various xAI techniques on a clinician's interaction with an AI-based DSS in decision-making tasks as compared to a general population. METHODS: We conducted a randomized, blinded study in which members of the Child Neurology Society and American Academy of Neurology were compared to a general population. Participants received recommendations from a DSS via a random assignment of an xAI intervention (decision tree, crowd sourced agreement, case-based reasoning, probability scores, counterfactual reasoning, feature importance, templated language, and no explanations). Primary outcomes included test performance and perceived explainability, trust, and social competence of the DSS. Secondary outcomes included compliance, understandability, and agreement per question. RESULTS: We had 81 neurology participants with 284 in the general population. Decision trees were perceived as the more explainable by the medical versus general population (P < 0.01) and as more explainable than probability scores within the medical population (P < 0.001). Increasing neurology experience and perceived explainability degraded performance (P = 0.0214). Performance was not predicted by xAI method but by perceived explainability. INTERPRETATION: xAI methods have different impacts on a medical versus general population; thus, xAI is not uniformly beneficial, and there is no one-size-fits-all approach. Further user-centered xAI research targeting clinicians and to develop personalized DSS for clinicians is needed.
目的:人工智能(AI)为基础的决策支持系统(DSS)在医学中得到了广泛应用,但决策过程背后的原理通常是未知的。可解释人工智能(xAI)技术为 DSS 提供了深入了解的途径,但对于如何为临床医生设计 xAI 技术,我们知之甚少。在这里,我们研究了与一般人群相比,各种 xAI 技术对临床医生在决策任务中与基于 AI 的 DSS 交互的影响。
方法:我们进行了一项随机、盲法研究,比较了儿童神经病学会和美国神经病学会成员与一般人群的差异。参与者通过 DSS 接收建议,通过 xAI 干预(决策树、众包共识、基于案例的推理、概率评分、反事实推理、特征重要性、模板语言和无解释)的随机分配来接收建议。主要结果包括测试表现以及对 DSS 的可解释性、信任和社会能力的感知。次要结果包括每个问题的合规性、可理解性和一致性。
结果:我们共有 81 名神经病学参与者和 284 名普通人群参与者。与一般人群相比,医学人群认为决策树比概率评分更具可解释性(P<0.01),也比一般人群更具可解释性(P<0.001)。增加神经病学经验和可感知的可解释性会降低表现(P=0.0214)。表现不是由 xAI 方法预测的,而是由感知的可解释性预测的。
解释:xAI 方法对医学人群和一般人群的影响不同;因此,xAI 并非普遍有益,也没有一刀切的方法。需要针对临床医生进行更多以用户为中心的 xAI 研究,并为临床医生开发个性化的 DSS。
Ann Clin Transl Neurol. 2024-5
2024 Int Conf Act Behav Comput (2024). 2024-5
BMC Med Ethics. 2024-10-1
Diagnostics (Basel). 2022-1-19
Crit Care. 2024-9-12
IEEE J Biomed Health Inform. 2024-7
Int J Med Inform. 2025-3
Patterns (N Y). 2024-6-14
Annu Rev Pathol. 2024-1-24
Med Image Anal. 2022-7
Lancet Digit Health. 2021-11
Kunstliche Intell (Oldenbourg). 2020
Wiley Interdiscip Rev Data Min Knowl Discov. 2019
Mol Inform. 2011-9-5