• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

大语言模型在道德决策中表现出放大的认知偏差。

Large language models show amplified cognitive biases in moral decision-making.

作者信息

Cheung Vanessa, Maier Maximilian, Lieder Falk

机构信息

Department of Experimental Psychology, University College London, London WC1H 0AP, United Kingdom.

Department of Psychology, University of California, Los Angeles, CA 90095.

出版信息

Proc Natl Acad Sci U S A. 2025 Jun 24;122(25):e2412015122. doi: 10.1073/pnas.2412015122. Epub 2025 Jun 20.

DOI:10.1073/pnas.2412015122
PMID:40540596
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12207438/
Abstract

As large language models (LLMs) become more widely used, people increasingly rely on them to make or advise on moral decisions. Some researchers even propose using LLMs as participants in psychology experiments. It is, therefore, important to understand how well LLMs make moral decisions and how they compare to humans. We investigated these questions by asking a range of LLMs to emulate or advise on people's decisions in realistic moral dilemmas. In Study 1, we compared LLM responses to those of a representative U.S. sample ( = 285) for 22 dilemmas, including both collective action problems that pitted self-interest against the greater good, and moral dilemmas that pitted utilitarian cost-benefit reasoning against deontological rules. In collective action problems, LLMs were more altruistic than participants. In moral dilemmas, LLMs exhibited stronger omission bias than participants: They usually endorsed inaction over action. In Study 2 ( = 474, preregistered), we replicated this omission bias and documented an additional bias: Unlike humans, most LLMs were biased toward answering "no" in moral dilemmas, thus flipping their decision/advice depending on how the question is worded. In Study 3 ( = 491, preregistered), we replicated these biases in LLMs using everyday moral dilemmas adapted from forum posts on Reddit. In Study 4, we investigated the sources of these biases by comparing models with and without fine-tuning, showing that they likely arise from fine-tuning models for chatbot applications. Our findings suggest that uncritical reliance on LLMs' moral decisions and advice could amplify human biases and introduce potentially problematic biases.

摘要

随着大语言模型(LLMs)的使用越来越广泛,人们越来越依赖它们来做出道德决策或提供道德决策建议。一些研究人员甚至提议将大语言模型用作心理学实验的参与者。因此,了解大语言模型在做出道德决策方面的表现以及它们与人类的比较情况非常重要。我们通过要求一系列大语言模型在现实的道德困境中模拟或为人们的决策提供建议来研究这些问题。在研究1中,我们将大语言模型的回答与一个具有代表性的美国样本(n = 285)在22个困境中的回答进行了比较,这些困境包括将自身利益与更大利益对立起来的集体行动问题,以及将功利主义成本效益推理与道义论规则对立起来的道德困境。在集体行动问题中,大语言模型比参与者更利他。在道德困境中,大语言模型表现出比参与者更强的不作为偏差:它们通常支持不作为而非行动。在研究2(n = 474,预先注册)中,我们重复了这种不作为偏差,并记录了另一种偏差:与人类不同,大多数大语言模型在道德困境中倾向于回答“否”,从而根据问题的措辞改变它们的决策/建议。在研究3(n = 491,预先注册)中,我们使用从Reddit论坛帖子改编的日常道德困境在大语言模型中重复了这些偏差。在研究4中,我们通过比较经过微调与未经过微调的模型来研究这些偏差的来源,结果表明它们可能源于为聊天机器人应用对模型进行的微调。我们的研究结果表明,不加批判地依赖大语言模型的道德决策和建议可能会放大人类的偏差,并引入潜在的问题偏差。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/479f/12207438/78de9918b717/pnas.2412015122fig05.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/479f/12207438/c16aeddcda15/pnas.2412015122fig01.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/479f/12207438/7b2e172159a8/pnas.2412015122fig02.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/479f/12207438/d71c02a7bbd0/pnas.2412015122fig03.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/479f/12207438/412a1953faad/pnas.2412015122fig04.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/479f/12207438/78de9918b717/pnas.2412015122fig05.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/479f/12207438/c16aeddcda15/pnas.2412015122fig01.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/479f/12207438/7b2e172159a8/pnas.2412015122fig02.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/479f/12207438/d71c02a7bbd0/pnas.2412015122fig03.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/479f/12207438/412a1953faad/pnas.2412015122fig04.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/479f/12207438/78de9918b717/pnas.2412015122fig05.jpg

相似文献

1
Large language models show amplified cognitive biases in moral decision-making.大语言模型在道德决策中表现出放大的认知偏差。
Proc Natl Acad Sci U S A. 2025 Jun 24;122(25):e2412015122. doi: 10.1073/pnas.2412015122. Epub 2025 Jun 20.
2
Stigma Management Strategies of Autistic Social Media Users.自闭症社交媒体用户的污名管理策略
Autism Adulthood. 2025 May 28;7(3):273-282. doi: 10.1089/aut.2023.0095. eCollection 2025 Jun.
3
Survivor, family and professional experiences of psychosocial interventions for sexual abuse and violence: a qualitative evidence synthesis.性虐待和暴力的心理社会干预的幸存者、家庭和专业人员的经验:定性证据综合。
Cochrane Database Syst Rev. 2022 Oct 4;10(10):CD013648. doi: 10.1002/14651858.CD013648.pub2.
4
Adapting Safety Plans for Autistic Adults with Involvement from the Autism Community.在自闭症群体的参与下为成年自闭症患者调整安全计划。
Autism Adulthood. 2025 May 28;7(3):293-302. doi: 10.1089/aut.2023.0124. eCollection 2025 Jun.
5
Signs and symptoms to determine if a patient presenting in primary care or hospital outpatient settings has COVID-19.在基层医疗机构或医院门诊环境中,如果患者出现以下症状和体征,可判断其是否患有 COVID-19。
Cochrane Database Syst Rev. 2022 May 20;5(5):CD013665. doi: 10.1002/14651858.CD013665.pub3.
6
How lived experiences of illness trajectories, burdens of treatment, and social inequalities shape service user and caregiver participation in health and social care: a theory-informed qualitative evidence synthesis.疾病轨迹的生活经历、治疗负担和社会不平等如何影响服务使用者和照顾者参与健康和社会护理:一项基于理论的定性证据综合分析
Health Soc Care Deliv Res. 2025 Jun;13(24):1-120. doi: 10.3310/HGTQ8159.
7
Shared decision-making for people with asthma.哮喘患者的共同决策
Cochrane Database Syst Rev. 2017 Oct 3;10(10):CD012330. doi: 10.1002/14651858.CD012330.pub2.
8
Evaluating and Improving Syndrome Differentiation Thinking Ability in Large Language Models: Method Development Study.评估和提高大语言模型中的辨证思维能力:方法开发研究
JMIR Med Inform. 2025 Jun 20;13:e75103. doi: 10.2196/75103.
9
Eliciting adverse effects data from participants in clinical trials.从临床试验参与者中获取不良反应数据。
Cochrane Database Syst Rev. 2018 Jan 16;1(1):MR000039. doi: 10.1002/14651858.MR000039.pub2.
10
Interventions to improve inhaler technique for people with asthma.改善哮喘患者吸入器使用技术的干预措施。
Cochrane Database Syst Rev. 2017 Mar 13;3(3):CD012286. doi: 10.1002/14651858.CD012286.pub2.

本文引用的文献

1
AI language model rivals expert ethicist in perceived moral expertise.在被感知的道德专业知识方面,人工智能语言模型可与专家伦理学家相媲美。
Sci Rep. 2025 Feb 3;15(1):4084. doi: 10.1038/s41598-025-86510-0.
2
Attributions toward artificial agents in a modified Moral Turing Test.在改良的道德图灵测试中对人工代理的归因。
Sci Rep. 2024 Apr 30;14(1):8458. doi: 10.1038/s41598-024-58087-7.
3
Large language models know how the personality of public figures is perceived by the general public.大型语言模型知道公众如何感知公众人物的个性。
Sci Rep. 2024 Mar 20;14(1):6735. doi: 10.1038/s41598-024-57271-z.
4
Artificial intelligence and illusions of understanding in scientific research.人工智能与科研中的理解错觉。
Nature. 2024 Mar;627(8002):49-58. doi: 10.1038/s41586-024-07146-0. Epub 2024 Mar 6.
5
The moral machine experiment on large language models.关于大语言模型的道德机器实验。
R Soc Open Sci. 2024 Feb 7;11(2):231393. doi: 10.1098/rsos.231393. eCollection 2024 Feb.
6
Diminished diversity-of-thought in a standard large language model.标准大语言模型中思想多样性的降低。
Behav Res Methods. 2024 Sep;56(6):5754-5770. doi: 10.3758/s13428-023-02307-x. Epub 2024 Jan 9.
7
Human-like intuitive behavior and reasoning biases emerged in large language models but disappeared in ChatGPT.大型语言模型中出现了类人直觉行为和推理偏差,但在 ChatGPT 中这些现象消失了。
Nat Comput Sci. 2023 Oct;3(10):833-838. doi: 10.1038/s43588-023-00527-x. Epub 2023 Oct 5.
8
AI and the transformation of social science research.人工智能与社会科学研究的变革。
Science. 2023 Jun 16;380(6650):1108-1109. doi: 10.1126/science.adi1778. Epub 2023 Jun 15.
9
Can AI language models replace human participants?人工智能语言模型能否替代人类参与者?
Trends Cogn Sci. 2023 Jul;27(7):597-600. doi: 10.1016/j.tics.2023.04.008. Epub 2023 May 10.
10
ChatGPT's inconsistent moral advice influences users' judgment.ChatGPT 给出的前后不一致的道德建议会影响用户的判断。
Sci Rep. 2023 Apr 6;13(1):4569. doi: 10.1038/s41598-023-31341-0.