Colombatto Clara, Birch Jonathan, Fleming Stephen M
Department of Psychology, University of Waterloo, Waterloo, ON, Canada.
Department of Philosophy, Logic and Scientific Method, and Centre for Philosophy of Natural and Social Science, London School of Economics and Political Science, London, UK.
Commun Psychol. 2025 May 25;3(1):84. doi: 10.1038/s44271-025-00262-1.
Rapid advances in artificial intelligence (AI) have led users to believe that systems such as large language models (LLMs) have mental states, including the capacity for 'experience' (e.g., emotions and consciousness). These folk-psychological attributions often diverge from expert opinion and are distinct from attributions of 'intelligence' (e.g., reasoning, planning), and yet may affect trust in AI systems. While past work provides some support for a link between anthropomorphism and trust, the impact of attributions of consciousness and other aspects of mentality on user trust remains unclear. We explored this in a preregistered experiment (N = 410) in which participants rated the capacity of an LLM to exhibit consciousness and a variety of other mental states. They then completed a decision-making task where they could revise their choices based on the advice of an LLM. Bayesian analyses revealed strong evidence against a positive correlation between attributions of consciousness and advice-taking; indeed, a dimension of mental states related to experience showed a negative relationship with advice-taking, while attributions of intelligence were strongly correlated with advice acceptance. These findings highlight how users' attitudes and behaviours are shaped by sophisticated intuitions about the capacities of LLMs-with different aspects of mental state attribution predicting people's trust in these systems.
人工智能(AI)的迅速发展让用户认为,诸如大语言模型(LLM)之类的系统具有心理状态,包括“体验”能力(如情感和意识)。这些民间心理学归因往往与专家意见不同,也有别于“智能”归因(如推理、规划),但可能会影响对人工智能系统的信任。虽然过去的研究为拟人化与信任之间的联系提供了一些支持,但意识归因和心理其他方面对用户信任的影响仍不明确。我们在一项预先注册的实验(N = 410)中对此进行了探究,在该实验中,参与者对一个大语言模型展现意识和各种其他心理状态的能力进行了评分。然后,他们完成了一项决策任务,在该任务中他们可以根据大语言模型的建议修改自己的选择。贝叶斯分析揭示了有力证据,反对意识归因与接受建议之间存在正相关;事实上,与体验相关的心理状态维度与接受建议呈负相关,而智能归因与接受建议密切相关。这些发现凸显了用户的态度和行为是如何受到对大语言模型能力的复杂直觉影响的——心理状态归因的不同方面预测了人们对这些系统的信任。