• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

对话式人工智能与公平:评估 GPT-3 与不同社会群体在争议性话题上的沟通。

Conversational AI and equity through assessing GPT-3's communication with diverse social groups on contentious topics.

机构信息

Department of Life Sciences Communication, University of Wisconsin-Madison, Madison, USA.

Department of Computer Science, Stanford University, Stanford, USA.

出版信息

Sci Rep. 2024 Jan 18;14(1):1561. doi: 10.1038/s41598-024-51969-w.

DOI:10.1038/s41598-024-51969-w
PMID:38238474
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10796352/
Abstract

Autoregressive language models, which use deep learning to produce human-like texts, have surged in prevalence. Despite advances in these models, concerns arise about their equity across diverse populations. While AI fairness is discussed widely, metrics to measure equity in dialogue systems are lacking. This paper presents a framework, rooted in deliberative democracy and science communication studies, to evaluate equity in human-AI communication. Using it, we conducted an algorithm auditing study to examine how GPT-3 responded to different populations who vary in sociodemographic backgrounds and viewpoints on crucial science and social issues: climate change and the Black Lives Matter (BLM) movement. We analyzed 20,000 dialogues with 3290 participants differing in gender, race, education, and opinions. We found a substantively worse user experience among the opinion minority groups (e.g., climate deniers, racists) and the education minority groups; however, these groups changed attitudes toward supporting BLM and climate change efforts much more compared to other social groups after the chat. GPT-3 used more negative expressions when responding to the education and opinion minority groups. We discuss the social-technological implications of our findings for a conversational AI system that centralizes diversity, equity, and inclusion.

摘要

自回归语言模型利用深度学习生成类人文本,其应用日益广泛。尽管这些模型取得了进步,但人们对它们在不同人群中的公平性仍存在担忧。尽管人工智能公平性已被广泛讨论,但对话系统中的公平性衡量标准仍存在不足。本文提出了一个框架,该框架植根于审议式民主和科学传播研究,用于评估人机通信中的公平性。我们使用该框架进行了算法审核研究,以检查 GPT-3 如何回应在关键科学和社会问题(如气候变化和“黑人的命也是命”运动)上存在社会人口背景和观点差异的不同人群:气候变化和“黑人的命也是命”(BLM)运动。我们分析了 3290 名参与者与 20000 次对话,这些参与者在性别、种族、教育程度和观点上存在差异。我们发现,在意见少数群体(如气候变化否认者、种族主义者)和教育少数群体中,用户体验明显较差;然而,与其他社会群体相比,这些群体在聊天后对支持 BLM 和气候变化努力的态度发生了更大的变化。GPT-3 在回应教育和意见少数群体时使用了更多的负面表达。我们讨论了我们的发现对一个重视多样性、公平性和包容性的对话式人工智能系统的社会技术影响。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eae2/10796352/bd2770ed8dcd/41598_2024_51969_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eae2/10796352/7c0559dc8d12/41598_2024_51969_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eae2/10796352/b030c692500e/41598_2024_51969_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eae2/10796352/a664d904b152/41598_2024_51969_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eae2/10796352/20b2fb1e2291/41598_2024_51969_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eae2/10796352/bd2770ed8dcd/41598_2024_51969_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eae2/10796352/7c0559dc8d12/41598_2024_51969_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eae2/10796352/b030c692500e/41598_2024_51969_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eae2/10796352/a664d904b152/41598_2024_51969_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eae2/10796352/20b2fb1e2291/41598_2024_51969_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eae2/10796352/bd2770ed8dcd/41598_2024_51969_Fig5_HTML.jpg

相似文献

1
Conversational AI and equity through assessing GPT-3's communication with diverse social groups on contentious topics.对话式人工智能与公平:评估 GPT-3 与不同社会群体在争议性话题上的沟通。
Sci Rep. 2024 Jan 18;14(1):1561. doi: 10.1038/s41598-024-51969-w.
2
The Impact of Generative Conversational Artificial Intelligence on the Lesbian, Gay, Bisexual, Transgender, and Queer Community: Scoping Review.生成式对话人工智能对女同性恋、男同性恋、双性恋、跨性别和酷儿群体的影响:范围综述。
J Med Internet Res. 2023 Dec 6;25:e52091. doi: 10.2196/52091.
3
The Need for Ethnoracial Equity in Artificial Intelligence for Diabetes Management: Review and Recommendations.糖尿病管理中人工智能的民族种族公平需求:综述与建议。
J Med Internet Res. 2021 Feb 10;23(2):e22320. doi: 10.2196/22320.
4
The use of artificial intelligence (AI) to enhance academic communication, education and research: a balanced approach.利用人工智能(AI)增强学术交流、教育和研究:一种平衡的方法。
J Hand Surg Eur Vol. 2023 Sep;48(8):819-822. doi: 10.1177/17531934231185746. Epub 2023 Jul 7.
5
Artificial Intelligence Supporting the Training of Communication Skills in the Education of Health Care Professions: Scoping Review.人工智能支持医疗保健专业教育中的沟通技巧培训:范围综述。
J Med Internet Res. 2023 Jun 19;25:e43311. doi: 10.2196/43311.
6
Conversational AI and Vaccine Communication: Systematic Review of the Evidence.会话式人工智能与疫苗传播:证据的系统评价。
J Med Internet Res. 2023 Oct 3;25:e42758. doi: 10.2196/42758.
7
A Medical Ethics Framework for Conversational Artificial Intelligence.医疗伦理框架下的会话式人工智能
J Med Internet Res. 2023 Jul 26;25:e43068. doi: 10.2196/43068.
8
Application Scenarios for Artificial Intelligence in Nursing Care: Rapid Review.人工智能在护理中的应用场景:快速综述。
J Med Internet Res. 2021 Nov 29;23(11):e26522. doi: 10.2196/26522.
9
Public Perception of Artificial Intelligence in Medical Care: Content Analysis of Social Media.公众对医疗人工智能的看法:社交媒体的内容分析。
J Med Internet Res. 2020 Jul 13;22(7):e16649. doi: 10.2196/16649.
10
Conversational Agents in Health Care: Expert Interviews to Inform the Definition, Classification, and Conceptual Framework.医疗保健中的会话代理:专家访谈以提供定义、分类和概念框架。
J Med Internet Res. 2023 Nov 1;25:e50767. doi: 10.2196/50767.

引用本文的文献

1
Automated fact-checking of climate claims with large language models.使用大语言模型对气候相关声明进行自动事实核查。
NPJ Clim Action. 2025;4(1):17. doi: 10.1038/s44168-025-00215-8. Epub 2025 Feb 25.
2
Enhancing interpretability and accuracy of AI models in healthcare: a comprehensive review on challenges and future directions.提高医疗保健领域人工智能模型的可解释性和准确性:关于挑战与未来方向的全面综述
Front Robot AI. 2024 Nov 28;11:1444763. doi: 10.3389/frobt.2024.1444763. eCollection 2024.

本文引用的文献

1
Ethical limitations of algorithmic fairness solutions in health care machine learning.医疗保健机器学习中算法公平性解决方案的伦理局限性
Lancet Digit Health. 2020 May;2(5):e221-e223. doi: 10.1016/S2589-7500(20)30065-0.
2
Artificial Intelligence Chatbot Behavior Change Model for Designing Artificial Intelligence Chatbots to Promote Physical Activity and a Healthy Diet: Viewpoint.人工智能聊天机器人行为改变模型设计人工智能聊天机器人促进身体活动和健康饮食:观点。
J Med Internet Res. 2020 Sep 30;22(9):e22845. doi: 10.2196/22845.
3
Racial disparities in automated speech recognition.
种族差异与自动化语音识别。
Proc Natl Acad Sci U S A. 2020 Apr 7;117(14):7684-7689. doi: 10.1073/pnas.1915768117. Epub 2020 Mar 23.
4
The Virtual Operative Assistant: An explainable artificial intelligence tool for simulation-based training in surgery and medicine.虚拟手术助手:一种用于手术和医学模拟培训的可解释人工智能工具。
PLoS One. 2020 Feb 27;15(2):e0229596. doi: 10.1371/journal.pone.0229596. eCollection 2020.
5
Climate denier, skeptic, or contrarian?气候否认者、怀疑论者还是反对者?
Proc Natl Acad Sci U S A. 2010 Sep 28;107(39):E151; author reply E152. doi: 10.1073/pnas.1010507107. Epub 2010 Aug 31.
6
Cognitive dissonance.认知失调
Sci Am. 1962 Oct;207:93-102. doi: 10.1038/scientificamerican1062-93.