Suppr超能文献

在聊天机器人兴起的背景下,对人工智能系统判断道德和能力的资质的认知。

Perceptions of artificial intelligence system's aptitude to judge morality and competence amidst the rise of Chatbots.

作者信息

Oliveira Manuel, Brands Justus, Mashudi Judith, Liefooghe Baptist, Hortensius Ruud

机构信息

Department of Industrial Engineering and Innovation Sciences, Eindhoven University of Technology, Eindhoven, The Netherlands.

Department of Psychology, Utrecht University, Utrecht, The Netherlands.

出版信息

Cogn Res Princ Implic. 2024 Jul 18;9(1):47. doi: 10.1186/s41235-024-00573-7.

Abstract

This paper examines how humans judge the capabilities of artificial intelligence (AI) to evaluate human attributes, specifically focusing on two key dimensions of human social evaluation: morality and competence. Furthermore, it investigates the impact of exposure to advanced Large Language Models on these perceptions. In three studies (combined N = 200), we tested the hypothesis that people will find it less plausible that AI is capable of judging the morality conveyed by a behavior compared to judging its competence. Participants estimated the plausibility of AI origin for a set of written impressions of positive and negative behaviors related to morality and competence. Studies 1 and 3 supported our hypothesis that people would be more inclined to attribute AI origin to competence-related impressions compared to morality-related ones. In Study 2, we found this effect only for impressions of positive behaviors. Additional exploratory analyses clarified that the differentiation between the AI origin of competence and morality judgments persisted throughout the first half year after the public launch of popular AI chatbot (i.e., ChatGPT) and could not be explained by participants' general attitudes toward AI, or the actual source of the impressions (i.e., AI or human). These findings suggest an enduring belief that AI is less adept at assessing the morality compared to the competence of human behavior, even as AI capabilities continued to advance.

摘要

本文探讨了人类如何判断人工智能(AI)评估人类属性的能力,特别关注人类社会评价的两个关键维度:道德和能力。此外,它还研究了接触先进的大语言模型对这些认知的影响。在三项研究(总样本量N = 200)中,我们检验了以下假设:与判断人类行为的能力相比,人们会觉得人工智能判断行为所传达的道德性的可信度较低。参与者估计了一系列与道德和能力相关的积极和消极行为的书面描述出自人工智能的可信度。研究1和研究3支持了我们的假设,即与道德相关的描述相比,人们更倾向于将人工智能来源归因于与能力相关的描述。在研究2中,我们发现这种效应仅适用于积极行为的描述。进一步的探索性分析表明,在流行的人工智能聊天机器人(即ChatGPT)公开发布后的上半年,能力判断和道德判断的人工智能来源之间的差异一直存在,并且不能用参与者对人工智能的总体态度或描述的实际来源(即人工智能或人类)来解释。这些发现表明,即使人工智能能力不断进步,人们仍然长期认为人工智能在评估人类行为的道德性方面不如评估能力那样擅长。

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验