• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

消除大语言模型的偏差:研究机遇。

Debiasing large language models: research opportunities.

作者信息

Yogarajan Vithya, Dobbie Gillian, Keegan Te Taka

机构信息

School of Computer Science, University of Auckland, Auckland, New Zealand.

School of Computing and Mathematical Sciences, University of Waikato, Hamilton, New Zealand.

出版信息

J R Soc N Z. 2024 Sep 16;55(2):372-395. doi: 10.1080/03036758.2024.2398567. eCollection 2025.

DOI:10.1080/03036758.2024.2398567
PMID:39677375
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11639098/
Abstract

Large language models (LLMs) are powerful decision-making tools widely adopted in healthcare, finance, and transportation. Embracing the opportunities and innovations of LLMs is inevitable. However, LLMs inherit stereotypes, misrepresentations, discrimination, and societies' biases from various sources-including training data, algorithm design, and user interactions-resulting in concerns about equality, diversity, and fairness. The bias problem has triggered increased research towards defining, detecting and quantifying bias and developing debiasing techniques. The predominant focus in tackling the bias problem is skewed towards resource-rich regions such as the US and Europe, resulting in a scarcity of research in other societies. As a small country with a unique history, culture and social composition, there is an opportunity for Aotearoa New Zealand's (NZ) research community to address this inadequacy. This paper presents an experimental evaluation of existing bias metrics and debiasing techniques in the NZ context. Research gaps derived from the study and a literature review are outlined, current and ongoing research in this space are discussed, and the suggested scope of research opportunities for NZ are presented.

摘要

大型语言模型(LLMs)是在医疗保健、金融和交通领域广泛采用的强大决策工具。拥抱大型语言模型带来的机遇和创新是不可避免的。然而,大型语言模型从包括训练数据、算法设计和用户交互在内的各种来源继承了刻板印象、错误表述、歧视和社会偏见,引发了对平等、多样性和公平性的担忧。偏差问题引发了对定义、检测和量化偏差以及开发去偏技术的研究增加。解决偏差问题的主要重点偏向于美国和欧洲等资源丰富的地区,导致其他社会的研究稀缺。作为一个拥有独特历史、文化和社会构成的小国,新西兰(NZ)的研究界有机会解决这一不足。本文对新西兰背景下现有的偏差度量和去偏技术进行了实验评估。概述了该研究和文献综述得出的研究差距,讨论了该领域当前和正在进行的研究,并提出了新西兰建议的研究机会范围。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/93aa/11639098/28930899983a/TNZR_A_2398567_F0005_OC.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/93aa/11639098/349eaffa82ee/TNZR_A_2398567_F0001_OB.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/93aa/11639098/b5858e1113ca/TNZR_A_2398567_F0002_OC.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/93aa/11639098/537438649f75/TNZR_A_2398567_F0003_OC.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/93aa/11639098/082ba6ea8f5b/TNZR_A_2398567_F0004_OC.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/93aa/11639098/28930899983a/TNZR_A_2398567_F0005_OC.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/93aa/11639098/349eaffa82ee/TNZR_A_2398567_F0001_OB.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/93aa/11639098/b5858e1113ca/TNZR_A_2398567_F0002_OC.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/93aa/11639098/537438649f75/TNZR_A_2398567_F0003_OC.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/93aa/11639098/082ba6ea8f5b/TNZR_A_2398567_F0004_OC.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/93aa/11639098/28930899983a/TNZR_A_2398567_F0005_OC.jpg

相似文献

1
Debiasing large language models: research opportunities.消除大语言模型的偏差:研究机遇。
J R Soc N Z. 2024 Sep 16;55(2):372-395. doi: 10.1080/03036758.2024.2398567. eCollection 2025.
2
Utilizing large language models for gastroenterology research: a conceptual framework.利用大语言模型进行胃肠病学研究:一个概念框架。
Therap Adv Gastroenterol. 2025 Apr 1;18:17562848251328577. doi: 10.1177/17562848251328577. eCollection 2025.
3
The Role of Large Language Models in Transforming Emergency Medicine: Scoping Review.大型语言模型在变革急诊医学中的作用:范围综述
JMIR Med Inform. 2024 May 10;12:e53787. doi: 10.2196/53787.
4
Fighting reviewer fatigue or amplifying bias? Considerations and recommendations for use of ChatGPT and other large language models in scholarly peer review.对抗审稿人疲劳还是加剧偏见?关于在学术同行评审中使用ChatGPT和其他大语言模型的思考与建议。
Res Integr Peer Rev. 2023 May 18;8(1):4. doi: 10.1186/s41073-023-00133-5.
5
Assessing the Alignment of Large Language Models With Human Values for Mental Health Integration: Cross-Sectional Study Using Schwartz's Theory of Basic Values.评估大型语言模型与人类心理健康整合价值观的一致性:使用施瓦茨基本价值观理论的横断面研究。
JMIR Ment Health. 2024 Apr 9;11:e55988. doi: 10.2196/55988.
6
Exploring Biases of Large Language Models in the Field of Mental Health: Comparative Questionnaire Study of the Effect of Gender and Sexual Orientation in Anorexia Nervosa and Bulimia Nervosa Case Vignettes.探索大语言模型在心理健康领域的偏差:神经性厌食症和神经性贪食症病例 vignettes 中性别和性取向影响的比较问卷调查研究。
JMIR Ment Health. 2025 Mar 20;12:e57986. doi: 10.2196/57986.
7
A Systematic Review of ChatGPT and Other Conversational Large Language Models in Healthcare.ChatGPT及其他对话式大语言模型在医疗保健领域的系统评价
medRxiv. 2024 Apr 27:2024.04.26.24306390. doi: 10.1101/2024.04.26.24306390.
8
Explicitly unbiased large language models still form biased associations.明确无偏见的大语言模型仍然会形成有偏见的关联。
Proc Natl Acad Sci U S A. 2025 Feb 25;122(8):e2416228122. doi: 10.1073/pnas.2416228122. Epub 2025 Feb 20.
9
DeepSeek in Healthcare: Revealing Opportunities and Steering Challenges of a New Open-Source Artificial Intelligence Frontier.医疗保健领域的DeepSeek:揭示新开源人工智能前沿的机遇与导向挑战
Cureus. 2025 Feb 18;17(2):e79221. doi: 10.7759/cureus.79221. eCollection 2025 Feb.
10
A survey of recent methods for addressing AI fairness and bias in biomedicine.生物医学中解决人工智能公平性和偏见问题的最新方法综述。
J Biomed Inform. 2024 Jun;154:104646. doi: 10.1016/j.jbi.2024.104646. Epub 2024 Apr 25.

引用本文的文献

1
Delving into the Practical Applications and Pitfalls of Large Language Models in Medical Education: Narrative Review.深入探讨大语言模型在医学教育中的实际应用与陷阱:叙述性综述
Adv Med Educ Pract. 2025 Apr 18;16:625-636. doi: 10.2147/AMEP.S497020. eCollection 2025.

本文引用的文献

1
Lessons learned from developing a COVID-19 algorithm governance framework in Aotearoa New Zealand.从新西兰制定新冠病毒算法治理框架中吸取的经验教训。
J R Soc N Z. 2022 Sep 19;53(1):82-94. doi: 10.1080/03036758.2022.2121290. eCollection 2023.
2
Holistic Evaluation of Language Models.语言模型的整体评估。
Ann N Y Acad Sci. 2023 Jul;1525(1):140-146. doi: 10.1111/nyas.15007. Epub 2023 May 25.
3
Stop Explaining Black Box Machine Learning Models for High Stakes Decisions and Use Interpretable Models Instead.停止为高风险决策解释黑箱机器学习模型,转而使用可解释模型。
Nat Mach Intell. 2019 May;1(5):206-215. doi: 10.1038/s42256-019-0048-x. Epub 2019 May 13.
4
Social bias, discrimination and inequity in healthcare: mechanisms, implications and recommendations.医疗保健中的社会偏见、歧视与不平等:机制、影响及建议
BJA Educ. 2022 Apr;22(4):131-137. doi: 10.1016/j.bjae.2021.11.011. Epub 2022 Feb 3.
5
Modern Clinical Text Mining: A Guide and Review.现代临床文本挖掘:指南与综述。
Annu Rev Biomed Data Sci. 2021 Jul 20;4:165-187. doi: 10.1146/annurev-biodatasci-030421-030931. Epub 2021 May 26.
6
Med-BERT: pretrained contextualized embeddings on large-scale structured electronic health records for disease prediction.医学BERT:基于大规模结构化电子健康记录进行疾病预测的预训练上下文嵌入模型
NPJ Digit Med. 2021 May 20;4(1):86. doi: 10.1038/s41746-021-00455-y.
7
Language models are an effective representation learning technique for electronic health record data.语言模型是一种用于电子健康记录数据的有效表示学习技术。
J Biomed Inform. 2021 Jan;113:103637. doi: 10.1016/j.jbi.2020.103637. Epub 2020 Dec 5.
8
BEHRT: Transformer for Electronic Health Records.BEHRT:电子健康记录的转换器。
Sci Rep. 2020 Apr 28;10(1):7155. doi: 10.1038/s41598-020-62922-y.
9
Racial disparities in automated speech recognition.种族差异与自动化语音识别。
Proc Natl Acad Sci U S A. 2020 Apr 7;117(14):7684-7689. doi: 10.1073/pnas.1915768117. Epub 2020 Mar 23.
10
Why cultural safety rather than cultural competency is required to achieve health equity: a literature review and recommended definition.为何实现健康公平需要文化安全而非文化能力:文献回顾与推荐定义
Int J Equity Health. 2019 Nov 14;18(1):174. doi: 10.1186/s12939-019-1082-3.