• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

我与机器?对人类和人工智能生成的建议的主观评估。

Me vs. the machine? Subjective evaluations of human- and AI-generated advice.

作者信息

Osborne Merrick R, Bailey Erica R

机构信息

U.C. Berkeley, Haas School of Business, Berkeley, United States.

出版信息

Sci Rep. 2025 Feb 1;15(1):3980. doi: 10.1038/s41598-025-86623-6.

DOI:10.1038/s41598-025-86623-6
PMID:39893236
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11787321/
Abstract

Artificial intelligence ("AI") has the potential to vastly improve human decision-making. In line with this, researchers have increasingly sought to understand how people view AI, often documenting skepticism and even outright aversion to these tools. In the present research, we complement these findings by documenting the performance of LLMs in the personal advice domain. In addition, we shift the focus in a new direction-exploring how interacting with AI tools, specifically large language models, impacts the user's view of themselves. In five preregistered experiments (N = 1,722), we explore evaluations of human- and ChatGPT-generated advice along three dimensions: quality, effectiveness, and authenticity. We find that ChatGPT produces superior advice relative to the average online participant even in a domain in which people strongly prefer human-generated advice (dating and relationships). We also document a bias against ChatGPT-generated advice which is present only when participants are aware the advice was generated by ChatGPT. Novel to the present investigation, we then explore how interacting with these tools impacts self-evaluations. We manipulate the order in which people interact with these tools relative to self-generation and find that generating advice before interacting with ChatGPT advice boosts the quality ratings of the ChatGPT advice. At the same time, interacting with ChatGPT-generated advice before self-generating advice decreases self-ratings of authenticity. Taken together, we document a bias towards AI in the context of personal advice. Further, we identify an important externality in the use of these tools-they can invoke social comparisons of me vs. the machine.

摘要

人工智能(“AI”)有潜力极大地改善人类决策。与此一致的是,研究人员越来越多地试图了解人们如何看待人工智能,经常记录下对这些工具的怀疑甚至完全厌恶。在本研究中,我们通过记录大语言模型在个人建议领域的表现来补充这些发现。此外,我们将重点转向一个新方向——探索与人工智能工具,特别是大语言模型的交互如何影响用户对自己的看法。在五个预先注册的实验(N = 1722)中,我们从三个维度探索对人类生成和ChatGPT生成的建议的评价:质量、有效性和真实性。我们发现,即使在人们强烈倾向于人类生成的建议的领域(约会和人际关系),ChatGPT生成的建议相对于普通在线参与者而言也更优。我们还记录了对ChatGPT生成的建议的偏见,这种偏见仅在参与者意识到建议是由ChatGPT生成时才会出现。本研究的新颖之处在于,我们随后探索了与这些工具的交互如何影响自我评估。我们操纵人们与这些工具交互相对于自我生成的顺序,发现先与ChatGPT的建议交互再自我生成建议会提高对ChatGPT建议的质量评级。与此同时,在自我生成建议之前与ChatGPT生成的建议交互会降低对真实性的自我评级。综上所述,我们记录了在个人建议背景下对人工智能的偏见。此外,我们识别出使用这些工具时一个重要的外部效应——它们会引发我与机器的社会比较。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a80b/11787321/8bfac8864c9a/41598_2025_86623_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a80b/11787321/f5311c1253f9/41598_2025_86623_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a80b/11787321/53fe496f9215/41598_2025_86623_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a80b/11787321/8bfac8864c9a/41598_2025_86623_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a80b/11787321/f5311c1253f9/41598_2025_86623_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a80b/11787321/53fe496f9215/41598_2025_86623_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a80b/11787321/8bfac8864c9a/41598_2025_86623_Fig3_HTML.jpg

相似文献

1
Me vs. the machine? Subjective evaluations of human- and AI-generated advice.我与机器?对人类和人工智能生成的建议的主观评估。
Sci Rep. 2025 Feb 1;15(1):3980. doi: 10.1038/s41598-025-86623-6.
2
Inconsistent advice by ChatGPT influences decision making in various areas.ChatGPT 提供的不一致建议会影响各个领域的决策。
Sci Rep. 2024 Jul 10;14(1):15876. doi: 10.1038/s41598-024-66821-4.
3
PICOT questions and search strategies formulation: A novel approach using artificial intelligence automation.PICOT问题与检索策略制定:一种使用人工智能自动化的新方法。
J Nurs Scholarsh. 2025 Jan;57(1):5-16. doi: 10.1111/jnu.13036. Epub 2024 Nov 24.
4
Between human and AI: assessing the reliability of AI text detection tools.在人与 AI 之间:评估 AI 文本检测工具的可靠性。
Curr Med Res Opin. 2024 Mar;40(3):353-358. doi: 10.1080/03007995.2024.2310086. Epub 2024 Feb 2.
5
Detecting Artificial Intelligence-Generated Versus Human-Written Medical Student Essays: Semirandomized Controlled Study.检测人工智能生成的与人类撰写的医学生论文:半随机对照研究。
JMIR Med Educ. 2025 Mar 3;11:e62779. doi: 10.2196/62779.
6
People devalue generative AI's competence but not its advice in addressing societal and personal challenges.人们贬低生成式人工智能应对社会和个人挑战的能力,但并不贬低其给出的建议。
Commun Psychol. 2023 Nov 15;1(1):32. doi: 10.1038/s44271-023-00032-x.
7
Artificial intelligence-large language models (AI-LLMs) for reliable and accurate cardiotocography (CTG) interpretation in obstetric practice.用于产科实践中可靠且准确解读胎心监护(CTG)的人工智能大语言模型(AI-LLMs)。
Comput Struct Biotechnol J. 2025 Mar 18;27:1140-1147. doi: 10.1016/j.csbj.2025.03.026. eCollection 2025.
8
Exploring artificial intelligence literacy and the use of ChatGPT and copilot in instruction on nursing academic report writing.探索人工智能素养以及ChatGPT和Copilot在护理学术报告写作教学中的应用。
Nurse Educ Today. 2025 Apr;147:106570. doi: 10.1016/j.nedt.2025.106570. Epub 2025 Jan 14.
9
Proficiency, Clarity, and Objectivity of Large Language Models Versus Specialists' Knowledge on COVID-19's Impacts in Pregnancy: Cross-Sectional Pilot Study.大型语言模型在新冠肺炎对妊娠影响方面的熟练度、清晰度和客观性与专家知识对比:横断面试点研究
JMIR Form Res. 2025 Feb 5;9:e56126. doi: 10.2196/56126.
10
Large Language Models for Therapy Recommendations Across 3 Clinical Specialties: Comparative Study.大型语言模型在 3 个临床专业领域的治疗推荐中的应用:比较研究。
J Med Internet Res. 2023 Oct 30;25:e49324. doi: 10.2196/49324.

引用本文的文献

1
Patient insights into empathy, compassion and self-disclosure in medical large language models: results from the IPALLM III study.患者对医学大语言模型中同理心、同情心和自我表露的见解:IPALLM III研究结果
World J Urol. 2025 Aug 14;43(1):492. doi: 10.1007/s00345-025-05872-2.

本文引用的文献

1
Bias against AI art can enhance perceptions of human creativity.对人工智能艺术的偏见可以增强人们对人类创造力的感知。
Sci Rep. 2023 Nov 3;13(1):19001. doi: 10.1038/s41598-023-45202-3.
2
Assessing the Utility of ChatGPT Throughout the Entire Clinical Workflow: Development and Usability Study.评估 ChatGPT 在整个临床工作流程中的效用:开发和可用性研究。
J Med Internet Res. 2023 Aug 22;25:e48659. doi: 10.2196/48659.
3
Comparing Physician and Artificial Intelligence Chatbot Responses to Patient Questions Posted to a Public Social Media Forum.
比较医生和人工智能聊天机器人对发布在公共社交媒体论坛上的患者问题的回复。
JAMA Intern Med. 2023 Jun 1;183(6):589-596. doi: 10.1001/jamainternmed.2023.1838.
4
Preference for human, not algorithm aversion.偏好人为,而非算法厌恶。
Trends Cogn Sci. 2022 Oct;26(10):824-826. doi: 10.1016/j.tics.2022.07.007. Epub 2022 Aug 5.
5
HUMAN DECISIONS AND MACHINE PREDICTIONS.人类决策与机器预测
Q J Econ. 2018 Feb 1;133(1):237-293. doi: 10.1093/qje/qjx032. Epub 2017 Aug 26.
6
Algorithm aversion: people erroneously avoid algorithms after seeing them err.算法厌恶:人们在看到算法出错后会错误地避免使用算法。
J Exp Psychol Gen. 2015 Feb;144(1):114-26. doi: 10.1037/xge0000033. Epub 2014 Nov 17.
7
The wisdom of select crowds.精选人群的智慧。
J Pers Soc Psychol. 2014 Aug;107(2):276-99. doi: 10.1037/a0036677.
8
When comparisons arise.当出现比较时。
J Pers Soc Psychol. 1995 Aug;69(2):227-36. doi: 10.1037//0022-3514.69.2.227.