• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

医学影像领域专家级视觉语言基础模型的人口统计学偏差

Demographic bias of expert-level vision-language foundation models in medical imaging.

作者信息

Yang Yuzhe, Liu Yujia, Liu Xin, Gulhane Avanti, Mastrodicasa Domenico, Wu Wei, Wang Edward J, Sahani Dushyant, Patel Shwetak

机构信息

Department of Electrical Engineering and Computer Science, Massachusetts Institute of Technology, Cambridge, MA, USA.

Department of Electrical and Computer Engineering, University of California, San Diego, La Jolla, CA, USA.

出版信息

Sci Adv. 2025 Mar 28;11(13):eadq0305. doi: 10.1126/sciadv.adq0305. Epub 2025 Mar 26.

DOI:10.1126/sciadv.adq0305
PMID:40138420
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11939055/
Abstract

Advances in artificial intelligence (AI) have achieved expert-level performance in medical imaging applications. Notably, self-supervised vision-language foundation models can detect a broad spectrum of pathologies without relying on explicit training annotations. However, it is crucial to ensure that these AI models do not mirror or amplify human biases, disadvantaging historically marginalized groups such as females or Black patients. In this study, we investigate the algorithmic fairness of state-of-the-art vision-language foundation models in chest x-ray diagnosis across five globally sourced datasets. Our findings reveal that compared to board-certified radiologists, these foundation models consistently underdiagnose marginalized groups, with even higher rates seen in intersectional subgroups such as Black female patients. Such biases present over a wide range of pathologies and demographic attributes. Further analysis of the model embedding uncovers its substantial encoding of demographic information. Deploying medical AI systems with biases can intensify preexisting care disparities, posing potential challenges to equitable healthcare access and raising ethical questions about their clinical applications.

摘要

人工智能(AI)的进展已在医学成像应用中达到了专家级水平。值得注意的是,自监督视觉语言基础模型无需依赖明确的训练标注就能检测出广泛的病症。然而,至关重要的是要确保这些人工智能模型不会反映或放大人类偏见,使女性或黑人患者等历史上处于边缘地位的群体处于不利地位。在本研究中,我们调查了跨五个全球来源数据集的胸部X光诊断中最先进的视觉语言基础模型的算法公平性。我们的研究结果表明,与获得委员会认证的放射科医生相比,这些基础模型始终对边缘化群体诊断不足,在黑人女性患者等交叉亚组中比例更高。这种偏见存在于广泛的病症和人口属性中。对模型嵌入的进一步分析揭示了其对人口信息的大量编码。部署存在偏见的医学人工智能系统可能会加剧现有的医疗差距,给公平获得医疗保健带来潜在挑战,并引发有关其临床应用的伦理问题。

相似文献

1
Demographic bias of expert-level vision-language foundation models in medical imaging.医学影像领域专家级视觉语言基础模型的人口统计学偏差
Sci Adv. 2025 Mar 28;11(13):eadq0305. doi: 10.1126/sciadv.adq0305. Epub 2025 Mar 26.
2
Underdiagnosis bias of artificial intelligence algorithms applied to chest radiographs in under-served patient populations.人工智能算法应用于服务不足患者人群的胸部 X 光片时的漏诊偏倚。
Nat Med. 2021 Dec;27(12):2176-2182. doi: 10.1038/s41591-021-01595-0. Epub 2021 Dec 10.
3
Fairness of artificial intelligence in healthcare: review and recommendations.人工智能在医疗保健中的公平性:综述与建议。
Jpn J Radiol. 2024 Jan;42(1):3-15. doi: 10.1007/s11604-023-01474-3. Epub 2023 Aug 4.
4
The limits of fair medical imaging AI in real-world generalization.公平的医学影像 AI 在现实世界泛化中的局限性。
Nat Med. 2024 Oct;30(10):2838-2848. doi: 10.1038/s41591-024-03113-4. Epub 2024 Jun 28.
5
Bias Mitigation in Primary Health Care Artificial Intelligence Models: Scoping Review.初级卫生保健人工智能模型中的偏差缓解:范围综述
J Med Internet Res. 2025 Jan 7;27:e60269. doi: 10.2196/60269.
6
Call for algorithmic fairness to mitigate amplification of racial biases in artificial intelligence models used in orthodontics and craniofacial health.呼吁算法公平性以减轻在口腔正畸学和颅面健康中使用的人工智能模型中种族偏见的放大。
Orthod Craniofac Res. 2023 Dec;26 Suppl 1:124-130. doi: 10.1111/ocr.12721. Epub 2023 Oct 17.
7
The imperative of diversity and equity for the adoption of responsible AI in healthcare.在医疗保健领域采用负责任的人工智能时,多样性和公平性的必要性。
Front Artif Intell. 2025 Apr 16;8:1577529. doi: 10.3389/frai.2025.1577529. eCollection 2025.
8
Evaluation and Bias Analysis of Large Language Models in Generating Synthetic Electronic Health Records: Comparative Study.大语言模型生成合成电子健康记录的评估与偏差分析:比较研究
J Med Internet Res. 2025 May 12;27:e65317. doi: 10.2196/65317.
9
Implications of Bias in Artificial Intelligence: Considerations for Cardiovascular Imaging.人工智能中的偏见问题及其对心血管成像的影响
Curr Atheroscler Rep. 2024 Apr;26(4):91-102. doi: 10.1007/s11883-024-01190-x. Epub 2024 Feb 16.
10
Artificial intelligence for breast cancer detection and its health technology assessment: A scoping review.用于乳腺癌检测的人工智能及其健康技术评估:一项范围综述。
Comput Biol Med. 2025 Jan;184:109391. doi: 10.1016/j.compbiomed.2024.109391. Epub 2024 Nov 22.

引用本文的文献

1
IHRAS: Automated Medical Report Generation from Chest X-Rays via Classification, Segmentation, and LLMs.IHRAS:通过分类、分割和大语言模型从胸部X光生成自动化医学报告
Bioengineering (Basel). 2025 Jul 24;12(8):795. doi: 10.3390/bioengineering12080795.
2
Artificial intelligence for cardiac imaging is ready for widespread clinical use: Pro Con debate AI for cardiac imaging.用于心脏成像的人工智能已准备好广泛应用于临床:支持与反对用于心脏成像的人工智能的辩论
BJR Open. 2025 Jun 6;7(1):tzaf015. doi: 10.1093/bjro/tzaf015. eCollection 2025 Jan.
3
Shortcut learning leads to sex bias in deep learning models for photoacoustic tomography.

本文引用的文献

1
The limits of fair medical imaging AI in real-world generalization.公平的医学影像 AI 在现实世界泛化中的局限性。
Nat Med. 2024 Oct;30(10):2838-2848. doi: 10.1038/s41591-024-03113-4. Epub 2024 Jun 28.
2
Medical Artificial Intelligence and Human Values.医学人工智能与人类价值观
N Engl J Med. 2024 May 30;390(20):1895-1904. doi: 10.1056/NEJMra2214183.
3
Demographic bias in misdiagnosis by computational pathology models.计算病理学模型导致的误诊中的人口统计学偏差。
捷径学习导致光声断层扫描深度学习模型中的性别偏见。
Int J Comput Assist Radiol Surg. 2025 May 9. doi: 10.1007/s11548-025-03370-9.
Nat Med. 2024 Apr;30(4):1174-1190. doi: 10.1038/s41591-024-02885-z. Epub 2024 Apr 19.
4
A visual-language foundation model for computational pathology.用于计算病理学的视觉-语言基础模型。
Nat Med. 2024 Mar;30(3):863-874. doi: 10.1038/s41591-024-02856-4. Epub 2024 Mar 19.
5
Risk of Bias in Chest Radiography Deep Learning Foundation Models.胸部X光深度学习基础模型中的偏倚风险
Radiol Artif Intell. 2023 Sep 27;5(6):e230060. doi: 10.1148/ryai.230060. eCollection 2023 Nov.
6
ExplaiNAble BioLogical Age (ENABL Age): an artificial intelligence framework for interpretable biological age.可解释生物学年龄(ENABL Age):一种用于可解释生物学年龄的人工智能框架。
Lancet Healthy Longev. 2023 Dec;4(12):e711-e723. doi: 10.1016/S2666-7568(23)00189-7. Epub 2023 Nov 6.
7
Improving model fairness in image-based computer-aided diagnosis.提高基于图像的计算机辅助诊断模型的公平性。
Nat Commun. 2023 Oct 6;14(1):6261. doi: 10.1038/s41467-023-41974-4.
8
Autonomous artificial intelligence increases real-world specialist clinic productivity in a cluster-randomized trial.在一项整群随机试验中,自主人工智能提高了现实世界专科诊所的工作效率。
NPJ Digit Med. 2023 Oct 4;6(1):184. doi: 10.1038/s41746-023-00931-7.
9
A foundation model for generalizable disease detection from retinal images.基于视网膜图像的通用疾病检测的基础模型。
Nature. 2023 Oct;622(7981):156-163. doi: 10.1038/s41586-023-06555-x. Epub 2023 Sep 13.
10
Using measures of race to make clinical predictions: Decision making, patient health, and fairness.种族指标在临床预测中的应用:决策、患者健康和公平性。
Proc Natl Acad Sci U S A. 2023 Aug 29;120(35):e2303370120. doi: 10.1073/pnas.2303370120. Epub 2023 Aug 22.