• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

大型语言模型是否表现出与人类相似的决策启发式?使用 GPT-3.5 的案例研究。

Do large language models show decision heuristics similar to humans? A case study using GPT-3.5.

机构信息

Department of Psychology, San Francisco State University.

出版信息

J Exp Psychol Gen. 2024 Apr;153(4):1066-1075. doi: 10.1037/xge0001547. Epub 2024 Feb 8.

DOI:10.1037/xge0001547
PMID:38330366
Abstract

A Large Language Model (LLM) is an artificial intelligence system trained on vast amounts of natural language data, enabling it to generate human-like responses to written or spoken language input. Generative Pre-Trained Transformer (GPT)-3.5 is an example of an LLM that supports a conversational agent called ChatGPT. In this work, we used a series of novel prompts to determine whether ChatGPT shows heuristics and other context-sensitive responses. We also tested the same prompts on human participants. Across four studies, we found that ChatGPT was influenced by random anchors in making estimates (anchoring, Study 1); it judged the likelihood of two events occurring together to be higher than the likelihood of either event occurring alone, and it was influenced by anecdotal information (representativeness and availability heuristic, Study 2); it found an item to be more efficacious when its features were presented positively rather than negatively-even though both presentations contained statistically equivalent information (framing effect, Study 3); and it valued an owned item more than a newly found item even though the two items were objectively identical (endowment effect, Study 4). In each study, human participants showed similar effects. Heuristics and context-sensitive responses in humans are thought to be driven by cognitive and affective processes such as loss aversion and effort reduction. The fact that an LLM-which lacks these processes-also shows such responses invites consideration of the possibility that language is sufficiently rich to carry these effects and may play a role in generating these effects in humans. (PsycInfo Database Record (c) 2024 APA, all rights reserved).

摘要

大型语言模型(LLM)是一种基于大量自然语言数据进行训练的人工智能系统,使其能够对书面或口头语言输入生成类似人类的响应。生成式预训练转换器(GPT)-3.5 是 LLM 的一个示例,它支持一个名为 ChatGPT 的对话代理。在这项工作中,我们使用了一系列新的提示来确定 ChatGPT 是否表现出启发式和其他上下文敏感的响应。我们还在人类参与者身上测试了相同的提示。在四项研究中,我们发现 ChatGPT 在进行估计时受到随机锚点的影响(锚定,研究 1);它判断两个事件同时发生的可能性高于任何一个事件单独发生的可能性,并且受到轶事信息的影响(代表性和可得性启发,研究 2);当它的特征以积极的方式而不是消极的方式呈现时,它会发现一个项目更有效,尽管两种呈现方式都包含了统计学上等效的信息(框架效应,研究 3);并且它更看重自己拥有的物品,而不是新发现的物品,尽管这两个物品在客观上是相同的(禀赋效应,研究 4)。在每项研究中,人类参与者都表现出了类似的效果。人类的启发式和上下文敏感响应被认为是由认知和情感过程驱动的,例如损失厌恶和减少努力。缺乏这些过程的 LLM 也表现出这种响应,这使得人们不得不考虑语言是否足够丰富,能够承载这些效果,并可能在人类中产生这些效果。(PsycInfo 数据库记录(c)2024 APA,保留所有权利)。

相似文献

1
Do large language models show decision heuristics similar to humans? A case study using GPT-3.5.大型语言模型是否表现出与人类相似的决策启发式?使用 GPT-3.5 的案例研究。
J Exp Psychol Gen. 2024 Apr;153(4):1066-1075. doi: 10.1037/xge0001547. Epub 2024 Feb 8.
2
Quality of Answers of Generative Large Language Models Versus Peer Users for Interpreting Laboratory Test Results for Lay Patients: Evaluation Study.生成式大语言模型与同行用户对解释非专业患者实验室检测结果的答案质量比较:评估研究。
J Med Internet Res. 2024 Apr 17;26:e56655. doi: 10.2196/56655.
3
Revolutionizing radiology with GPT-based models: Current applications, future possibilities and limitations of ChatGPT.基于 GPT 的模型推动放射学革命:ChatGPT 的当前应用、未来可能性和局限性。
Diagn Interv Imaging. 2023 Jun;104(6):269-274. doi: 10.1016/j.diii.2023.02.003. Epub 2023 Feb 28.
4
ChatGPT and large language model (LLM) chatbots: The current state of acceptability and a proposal for guidelines on utilization in academic medicine.ChatGPT 和大型语言模型 (LLM) 聊天机器人:在接受度方面的现状以及在学术医学中使用指南的建议。
J Pediatr Urol. 2023 Oct;19(5):598-604. doi: 10.1016/j.jpurol.2023.05.018. Epub 2023 Jun 2.
5
ChatGPT and Bard exhibit spontaneous citation fabrication during psychiatry literature search.ChatGPT 和 Bard 在搜索精神病学文献时会自发编造引文。
Psychiatry Res. 2023 Aug;326:115334. doi: 10.1016/j.psychres.2023.115334. Epub 2023 Jul 7.
6
Triage Performance Across Large Language Models, ChatGPT, and Untrained Doctors in Emergency Medicine: Comparative Study.分诊表现比较:大型语言模型、ChatGPT 和未经训练的急诊医生:一项对比研究。
J Med Internet Res. 2024 Jun 14;26:e53297. doi: 10.2196/53297.
7
Artificial Intelligence Can Generate Fraudulent but Authentic-Looking Scientific Medical Articles: Pandora's Box Has Been Opened.人工智能可以生成虚假但看起来真实的科学医学文章:潘多拉的盒子已经被打开。
J Med Internet Res. 2023 May 31;25:e46924. doi: 10.2196/46924.
8
How Does ChatGPT Perform on the United States Medical Licensing Examination (USMLE)? The Implications of Large Language Models for Medical Education and Knowledge Assessment.ChatGPT在美国医师执照考试(USMLE)中的表现如何?大语言模型对医学教育和知识评估的影响。
JMIR Med Educ. 2023 Feb 8;9:e45312. doi: 10.2196/45312.
9
A Language Model-Powered Simulated Patient With Automated Feedback for History Taking: Prospective Study.基于语言模型的模拟患者与自动化反馈的病史采集:前瞻性研究。
JMIR Med Educ. 2024 Aug 16;10:e59213. doi: 10.2196/59213.
10
A Comparative Study of Responses to Retina Questions from Either Experts, Expert-Edited Large Language Models, or Expert-Edited Large Language Models Alone.专家、经过专家编辑的大语言模型或仅经过专家编辑的大语言模型对视网膜问题回答的比较研究。
Ophthalmol Sci. 2024 Feb 6;4(4):100485. doi: 10.1016/j.xops.2024.100485. eCollection 2024 Jul-Aug.

引用本文的文献

1
Evaluating the ability of large Language models to predict human social decisions.评估大语言模型预测人类社会决策的能力。
Sci Rep. 2025 Sep 2;15(1):32290. doi: 10.1038/s41598-025-17188-7.
2
Will AI become our Co-PI?人工智能会成为我们的共同首席研究员吗?
NPJ Digit Med. 2025 Jul 14;8(1):440. doi: 10.1038/s41746-025-01859-w.
3
Using large language models to facilitate academic work in the psychological sciences.使用大语言模型促进心理科学领域的学术工作。
Curr Psychol. 2025;44(9):7910-7918. doi: 10.1007/s12144-025-07438-2. Epub 2025 Jan 28.
4
Kernels of selfhood: GPT-4o shows humanlike patterns of cognitive dissonance moderated by free choice.自我内核:GPT-4o展现出由自由选择调节的类似人类的认知失调模式。
Proc Natl Acad Sci U S A. 2025 May 20;122(20):e2501823122. doi: 10.1073/pnas.2501823122. Epub 2025 May 14.
5
Exploring ChatGPT's potential in the clinical stream of neurorehabilitation.探索ChatGPT在神经康复临床领域的潜力。
Front Artif Intell. 2024 Jun 6;7:1407905. doi: 10.3389/frai.2024.1407905. eCollection 2024.
6
Language models and psychological sciences.语言模型与心理科学。
Front Psychol. 2023 Oct 20;14:1279317. doi: 10.3389/fpsyg.2023.1279317. eCollection 2023.