Suppr超能文献

人们如何在熟悉和不熟悉的领域中,用反事实和因果解释来推理人工智能决策。

How people reason with counterfactual and causal explanations for Artificial Intelligence decisions in familiar and unfamiliar domains.

机构信息

School of Psychology and Institute of Neuroscience, Trinity College Dublin, University of Dublin, Dublin, Ireland.

出版信息

Mem Cognit. 2023 Oct;51(7):1481-1496. doi: 10.3758/s13421-023-01407-5. Epub 2023 Mar 24.

Abstract

Few empirical studies have examined how people understand counterfactual explanations for other people's decisions, for example, "if you had asked for a lower amount, your loan application would have been approved". Yet many current Artificial Intelligence (AI) decision support systems rely on counterfactual explanations to improve human understanding and trust. We compared counterfactual explanations to causal ones, i.e., "because you asked for a high amount, your loan application was not approved", for an AI's decisions in a familiar domain (alcohol and driving) and an unfamiliar one (chemical safety) in four experiments (n = 731). Participants were shown inputs to an AI system, its decisions, and an explanation for each decision; they attempted to predict the AI's decisions, or to make their own decisions. Participants judged counterfactual explanations more helpful than causal ones, but counterfactuals did not improve the accuracy of their predictions of the AI's decisions more than causals (Experiment 1). However, counterfactuals improved the accuracy of participants' own decisions more than causals (Experiment 2). When the AI's decisions were correct (Experiments 1 and 2), participants considered explanations more helpful and made more accurate judgements in the familiar domain than in the unfamiliar one; but when the AI's decisions were incorrect, they considered explanations less helpful and made fewer accurate judgements in the familiar domain than the unfamiliar one, whether they predicted the AI's decisions (Experiment 3a) or made their own decisions (Experiment 3b). The results corroborate the proposal that counterfactuals provide richer information than causals, because their mental representation includes more possibilities.

摘要

很少有实证研究考察人们如何理解他人决策的反事实解释,例如,“如果你要求的金额较低,你的贷款申请就会获得批准”。然而,许多当前的人工智能 (AI) 决策支持系统依赖于反事实解释来提高人类的理解和信任。我们在四个实验中比较了反事实解释和因果解释,即“因为你要求的金额较高,所以你的贷款申请未被批准”,用于 AI 在熟悉的领域(酒精和驾驶)和不熟悉的领域(化学安全)的决策(n = 731)。参与者被展示了 AI 系统的输入、决策及其每个决策的解释;他们试图预测 AI 的决策,或做出自己的决策。参与者认为反事实解释比因果解释更有帮助,但反事实并没有比因果解释更能提高他们对 AI 决策的预测准确性(实验 1)。然而,反事实解释比因果解释更能提高参与者自己决策的准确性(实验 2)。当 AI 的决策是正确的(实验 1 和 2)时,参与者认为在熟悉的领域比在不熟悉的领域中解释更有帮助,做出的判断更准确;但是当 AI 的决策是错误的时,他们认为在熟悉的领域比在不熟悉的领域中解释不那么有帮助,做出的判断也不那么准确,无论是预测 AI 的决策(实验 3a)还是做出自己的决策(实验 3b)。这些结果证实了这样的假设,即反事实比因果提供了更丰富的信息,因为它们的心理表示包括更多的可能性。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/babb/10520145/8770e9c320b0/13421_2023_1407_Fig1_HTML.jpg

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验