• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

减轻大语言模型加剧的健康不平等风险。

Mitigating the risk of health inequity exacerbated by large language models.

作者信息

Ji Yuelyu, Ma Wenhe, Sivarajkumar Sonish, Zhang Hang, Sadhu Eugene M, Li Zhuochun, Wu Xizhi, Visweswaran Shyam, Wang Yanshan

机构信息

Department of Information Science, University of Pittsburgh, Pittsburgh, PA, USA.

Department of Health Information Management, University of Pittsburgh, Pittsburgh, PA, USA.

出版信息

NPJ Digit Med. 2025 May 4;8(1):246. doi: 10.1038/s41746-025-01576-4.

DOI:10.1038/s41746-025-01576-4
PMID:40319154
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12049425/
Abstract

Recent advancements in large language models (LLMs) have demonstrated their potential in numerous medical applications, particularly in automating clinical trial matching for translational research and enhancing medical question-answering for clinical decision support. However, our study shows that incorporating non-decisive socio-demographic factors, such as race, sex, income level, LGBT+ status, homelessness, illiteracy, disability, and unemployment, into the input of LLMs can lead to incorrect and harmful outputs. These discrepancies could worsen existing health disparities if LLMs are broadly implemented in healthcare. To address this issue, we introduce EquityGuard, a novel framework designed to detect and mitigate the risk of health inequities in LLM-based medical applications. Our evaluation demonstrates its effectiveness in promoting equitable outcomes across diverse populations.

摘要

大语言模型(LLMs)的最新进展已在众多医学应用中展现出其潜力,尤其是在为转化研究自动进行临床试验匹配以及增强用于临床决策支持的医学问答方面。然而,我们的研究表明,将非决定性的社会人口因素,如种族、性别、收入水平、LGBT+身份、无家可归、文盲、残疾和失业等纳入大语言模型的输入,可能会导致错误且有害的输出。如果大语言模型在医疗保健领域广泛应用,这些差异可能会加剧现有的健康不平等。为解决这一问题,我们引入了EquityGuard,这是一个旨在检测和减轻基于大语言模型的医学应用中健康不平等风险的新颖框架。我们的评估证明了它在促进不同人群公平结果方面的有效性。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1ff0/12049425/875e149eb3c2/41746_2025_1576_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1ff0/12049425/1a451134285f/41746_2025_1576_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1ff0/12049425/e96454e21ab3/41746_2025_1576_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1ff0/12049425/b0b7961fd6b2/41746_2025_1576_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1ff0/12049425/71474411ffc8/41746_2025_1576_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1ff0/12049425/6e482ec2817e/41746_2025_1576_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1ff0/12049425/875e149eb3c2/41746_2025_1576_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1ff0/12049425/1a451134285f/41746_2025_1576_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1ff0/12049425/e96454e21ab3/41746_2025_1576_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1ff0/12049425/b0b7961fd6b2/41746_2025_1576_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1ff0/12049425/71474411ffc8/41746_2025_1576_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1ff0/12049425/6e482ec2817e/41746_2025_1576_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1ff0/12049425/875e149eb3c2/41746_2025_1576_Fig6_HTML.jpg

相似文献

1
Mitigating the risk of health inequity exacerbated by large language models.减轻大语言模型加剧的健康不平等风险。
NPJ Digit Med. 2025 May 4;8(1):246. doi: 10.1038/s41746-025-01576-4.
2
Performance of Large Language Models on a Neurology Board-Style Examination.大语言模型在神经科 board-style 考试中的表现。
JAMA Netw Open. 2023 Dec 1;6(12):e2346721. doi: 10.1001/jamanetworkopen.2023.46721.
3
The Role of Large Language Models in Transforming Emergency Medicine: Scoping Review.大型语言模型在变革急诊医学中的作用:范围综述
JMIR Med Inform. 2024 May 10;12:e53787. doi: 10.2196/53787.
4
Assessing and Optimizing Large Language Models on Spondyloarthritis Multi-Choice Question Answering: Protocol for Enhancement and Assessment.评估和优化用于脊柱关节炎多项选择题回答的大型语言模型:增强和评估的方案。
JMIR Res Protoc. 2024 May 24;13:e57001. doi: 10.2196/57001.
5
Large Language Models in Worldwide Medical Exams: Platform Development and Comprehensive Analysis.全球医学考试中的大语言模型:平台开发与综合分析
J Med Internet Res. 2024 Dec 27;26:e66114. doi: 10.2196/66114.
6
A dataset and benchmark for hospital course summarization with adapted large language models.一个用于医院病程总结的数据集和基准测试,采用了适配的大语言模型。
J Am Med Inform Assoc. 2025 Mar 1;32(3):470-479. doi: 10.1093/jamia/ocae312.
7
Large Language Models and User Trust: Consequence of Self-Referential Learning Loop and the Deskilling of Health Care Professionals.大语言模型与用户信任:自我参照学习循环的后果及医疗保健专业人员的技能退化
J Med Internet Res. 2024 Apr 25;26:e56764. doi: 10.2196/56764.
8
Large language models leverage external knowledge to extend clinical insight beyond language boundaries.大语言模型利用外部知识将临床洞察力扩展到语言边界之外。
J Am Med Inform Assoc. 2024 Sep 1;31(9):2054-2064. doi: 10.1093/jamia/ocae079.
9
One LLM is not Enough: Harnessing the Power of Ensemble Learning for Medical Question Answering.一个语言模型是不够的:利用集成学习的力量进行医学问答。
medRxiv. 2023 Dec 24:2023.12.21.23300380. doi: 10.1101/2023.12.21.23300380.
10
Benchmarking Vision Capabilities of Large Language Models in Surgical Examination Questions.大型语言模型在外科检查问题中的视觉能力基准测试
J Surg Educ. 2025 Apr;82(4):103442. doi: 10.1016/j.jsurg.2025.103442. Epub 2025 Feb 9.

引用本文的文献

1
Challenges of Implementing LLMs in Clinical Practice: Perspectives.在临床实践中应用大语言模型的挑战:观点
J Clin Med. 2025 Sep 1;14(17):6169. doi: 10.3390/jcm14176169.
2
Evaluating anti-LGBTQIA+ medical bias in large language models.评估大语言模型中针对 LGBTQIA+ 群体的医学偏见。
PLOS Digit Health. 2025 Sep 8;4(9):e0001001. doi: 10.1371/journal.pdig.0001001. eCollection 2025 Sep.
3
The Role of Large Language Models (LLMs) in Hepato-Pancreato-Biliary Surgery: Opportunities and Challenges.大语言模型在肝胰胆外科手术中的作用:机遇与挑战

本文引用的文献

1
Explicitly unbiased large language models still form biased associations.明确无偏见的大语言模型仍然会形成有偏见的关联。
Proc Natl Acad Sci U S A. 2025 Feb 25;122(8):e2416228122. doi: 10.1073/pnas.2416228122. Epub 2025 Feb 20.
2
Toward expert-level medical question answering with large language models.迈向使用大语言模型实现专家级医学问答
Nat Med. 2025 Mar;31(3):943-950. doi: 10.1038/s41591-024-03423-7. Epub 2025 Jan 8.
3
Matching patients to clinical trials with large language models.利用大型语言模型为患者匹配临床试验。
Cureus. 2025 Jun 14;17(6):e85979. doi: 10.7759/cureus.85979. eCollection 2025 Jun.
Nat Commun. 2024 Nov 18;15(1):9074. doi: 10.1038/s41467-024-53081-z.
4
Larger and more instructable language models become less reliable.更大且更具指导性的语言模型变得不那么可靠。
Nature. 2024 Oct;634(8032):61-68. doi: 10.1038/s41586-024-07930-y. Epub 2024 Sep 25.
5
Systematic evaluation of common natural language processing techniques to codify clinical notes.系统评估常见的自然语言处理技术以对临床记录进行编码。
PLoS One. 2024 Mar 7;19(3):e0298892. doi: 10.1371/journal.pone.0298892. eCollection 2024.
6
A novel attention-based cross-modal transfer learning framework for predicting cardiovascular disease.一种基于注意力的新型跨模态转移学习框架,用于预测心血管疾病。
Comput Biol Med. 2024 Mar;170:107977. doi: 10.1016/j.compbiomed.2024.107977. Epub 2024 Jan 9.
7
Leveraging Large Language Models for Decision Support in Personalized Oncology.利用大型语言模型为个性化肿瘤学提供决策支持。
JAMA Netw Open. 2023 Nov 1;6(11):e2343689. doi: 10.1001/jamanetworkopen.2023.43689.
8
Advancing AI in healthcare: A comprehensive review of best practices.推进医疗保健领域的人工智能:最佳实践的全面综述。
Clin Chim Acta. 2023 Aug 1;548:117519. doi: 10.1016/j.cca.2023.117519. Epub 2023 Aug 16.
9
Large language models encode clinical knowledge.大语言模型编码临床知识。
Nature. 2023 Aug;620(7972):172-180. doi: 10.1038/s41586-023-06291-2. Epub 2023 Jul 12.
10
Algorithmic fairness in artificial intelligence for medicine and healthcare.人工智能在医学和医疗保健中的算法公平性。
Nat Biomed Eng. 2023 Jun;7(6):719-742. doi: 10.1038/s41551-023-01056-8. Epub 2023 Jun 28.