• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

医学人工智能中的偏差:对临床决策的影响。

Bias in medical AI: Implications for clinical decision-making.

作者信息

Cross James L, Choma Michael A, Onofrey John A

机构信息

Yale School of Medicine, New Haven, Connecticut, United States of America.

Department of Radiology & Biomedical Imaging, Yale University, New Haven, Connecticut, United States of America.

出版信息

PLOS Digit Health. 2024 Nov 7;3(11):e0000651. doi: 10.1371/journal.pdig.0000651. eCollection 2024 Nov.

DOI:10.1371/journal.pdig.0000651
PMID:39509461
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11542778/
Abstract

Biases in medical artificial intelligence (AI) arise and compound throughout the AI lifecycle. These biases can have significant clinical consequences, especially in applications that involve clinical decision-making. Left unaddressed, biased medical AI can lead to substandard clinical decisions and the perpetuation and exacerbation of longstanding healthcare disparities. We discuss potential biases that can arise at different stages in the AI development pipeline and how they can affect AI algorithms and clinical decision-making. Bias can occur in data features and labels, model development and evaluation, deployment, and publication. Insufficient sample sizes for certain patient groups can result in suboptimal performance, algorithm underestimation, and clinically unmeaningful predictions. Missing patient findings can also produce biased model behavior, including capturable but nonrandomly missing data, such as diagnosis codes, and data that is not usually or not easily captured, such as social determinants of health. Expertly annotated labels used to train supervised learning models may reflect implicit cognitive biases or substandard care practices. Overreliance on performance metrics during model development may obscure bias and diminish a model's clinical utility. When applied to data outside the training cohort, model performance can deteriorate from previous validation and can do so differentially across subgroups. How end users interact with deployed solutions can introduce bias. Finally, where models are developed and published, and by whom, impacts the trajectories and priorities of future medical AI development. Solutions to mitigate bias must be implemented with care, which include the collection of large and diverse data sets, statistical debiasing methods, thorough model evaluation, emphasis on model interpretability, and standardized bias reporting and transparency requirements. Prior to real-world implementation in clinical settings, rigorous validation through clinical trials is critical to demonstrate unbiased application. Addressing biases across model development stages is crucial for ensuring all patients benefit equitably from the future of medical AI.

摘要

医学人工智能(AI)中的偏差在AI的整个生命周期中都会出现并不断累积。这些偏差可能会产生重大的临床后果,尤其是在涉及临床决策的应用中。如果不加以解决,有偏差的医学AI可能会导致临床决策不达标,以及长期存在的医疗保健差距持续存在并加剧。我们讨论了在AI开发流程的不同阶段可能出现的潜在偏差,以及它们如何影响AI算法和临床决策。偏差可能出现在数据特征和标签、模型开发与评估、部署以及发表等环节。某些患者群体的样本量不足可能导致性能欠佳、算法低估以及临床上无意义的预测。患者检查结果缺失也可能产生有偏差的模型行为,包括可获取但非随机缺失的数据,如诊断代码,以及通常不被获取或难以获取的数据,如健康的社会决定因素。用于训练监督学习模型的专业注释标签可能反映出隐性认知偏差或不规范的医疗实践。在模型开发过程中过度依赖性能指标可能会掩盖偏差并降低模型的临床效用。当应用于训练队列之外的数据时,模型性能可能会比之前的验证结果变差,并且在不同亚组中的表现可能存在差异。最终用户与已部署解决方案的交互方式可能会引入偏差。最后,模型的开发、发表地点以及开发者是谁,都会影响未来医学AI发展的轨迹和重点。减轻偏差的解决方案必须谨慎实施,这包括收集大量多样的数据集、统计去偏方法、全面的模型评估、强调模型可解释性,以及标准化的偏差报告和透明度要求。在临床环境中进行实际应用之前,通过临床试验进行严格验证对于证明无偏差应用至关重要。解决模型开发各阶段的偏差对于确保所有患者公平受益于医学AI的未来至关重要。

相似文献

1
Bias in medical AI: Implications for clinical decision-making.医学人工智能中的偏差:对临床决策的影响。
PLOS Digit Health. 2024 Nov 7;3(11):e0000651. doi: 10.1371/journal.pdig.0000651. eCollection 2024 Nov.
2
Leveraging code-free deep learning for pill recognition in clinical settings: A multicenter, real-world study of performance across multiple platforms.利用无代码深度学习在临床环境中进行药丸识别:在多个平台上进行的多中心真实世界性能研究。
Artif Intell Med. 2024 Apr;150:102844. doi: 10.1016/j.artmed.2024.102844. Epub 2024 Mar 13.
3
Unmasking bias in artificial intelligence: a systematic review of bias detection and mitigation strategies in electronic health record-based models.揭开人工智能中的偏见:基于电子健康记录模型的偏见检测和缓解策略的系统评价。
J Am Med Inform Assoc. 2024 Apr 19;31(5):1172-1183. doi: 10.1093/jamia/ocae060.
4
Implications of Bias in Artificial Intelligence: Considerations for Cardiovascular Imaging.人工智能中的偏见问题及其对心血管成像的影响
Curr Atheroscler Rep. 2024 Apr;26(4):91-102. doi: 10.1007/s11883-024-01190-x. Epub 2024 Feb 16.
5
Mitigating the impact of biased artificial intelligence in emergency decision-making.减轻有偏见的人工智能在应急决策中的影响。
Commun Med (Lond). 2022 Nov 21;2(1):149. doi: 10.1038/s43856-022-00214-4.
6
"Shortcuts" Causing Bias in Radiology Artificial Intelligence: Causes, Evaluation, and Mitigation.“捷径”导致放射科人工智能产生偏见:原因、评估和缓解。
J Am Coll Radiol. 2023 Sep;20(9):842-851. doi: 10.1016/j.jacr.2023.06.025. Epub 2023 Jul 27.
7
Implicit Bias隐性偏见
8
Challenges of artificial intelligence in medicine and dermatology.人工智能在医学和皮肤病学领域面临的挑战。
Clin Dermatol. 2024 May-Jun;42(3):210-215. doi: 10.1016/j.clindermatol.2023.12.013. Epub 2024 Jan 4.
9
A survey of recent methods for addressing AI fairness and bias in biomedicine.生物医学中解决人工智能公平性和偏见问题的最新方法综述。
J Biomed Inform. 2024 Jun;154:104646. doi: 10.1016/j.jbi.2024.104646. Epub 2024 Apr 25.
10
Unmasking bias in artificial intelligence: a systematic review of bias detection and mitigation strategies in electronic health record-based models.揭示人工智能中的偏见:基于电子健康记录模型的偏见检测与缓解策略的系统评价
ArXiv. 2024 Jul 1:arXiv:2310.19917v3.

引用本文的文献

1
Role of artificial intelligence in congenital heart disease.人工智能在先天性心脏病中的作用。
World J Clin Pediatr. 2025 Sep 9;14(3):105926. doi: 10.5409/wjcp.v14.i3.105926.
2
AI Methods Tailored to Influenza, RSV, HIV, and SARS-CoV-2: A Focused Review.针对流感、呼吸道合胞病毒、艾滋病毒和新型冠状病毒2的人工智能方法:重点综述
Pathogens. 2025 Jul 30;14(8):748. doi: 10.3390/pathogens14080748.
3
Transparent Machine Learning Reveals Diagnostic Glycan Biomarkers in Subarachnoid Hemorrhage and Vasospasm.透明机器学习揭示蛛网膜下腔出血和血管痉挛中的诊断性聚糖生物标志物。
Int J Mol Sci. 2025 Aug 10;26(16):7727. doi: 10.3390/ijms26167727.
4
Comparative Evaluation of Diagnosis and Treatment Plan Given by Pediatric Dentists and Generated by ChatGPT: A Cross-Sectional Pilot Study.儿科牙医给出的诊断和治疗方案与ChatGPT生成的诊断和治疗方案的比较评估:一项横断面试点研究。
Cureus. 2025 Jul 22;17(7):e88505. doi: 10.7759/cureus.88505. eCollection 2025 Jul.
5
Artificial intelligence in maxillofacial trauma: expert ally or unreliable assistant?人工智能在颌面创伤中的应用:专家助手还是不可靠的帮手?
Med Oral Patol Oral Cir Bucal. 2025 Sep 1;30(5):e751-e757. doi: 10.4317/medoral.27229.
6
Perceptions of large language models in medical education and clinical practice among pediatric emergency physicians in Saudi Arabia: a multiregional cross-sectional study.沙特阿拉伯儿科急诊医生对大语言模型在医学教育和临床实践中的认知:一项多地区横断面研究。
Front Public Health. 2025 Jul 30;13:1634638. doi: 10.3389/fpubh.2025.1634638. eCollection 2025.
7
Evaluation of Multiple-Choice Tests in Head and Neck Ultrasound Created by Physicians and Large Language Models.医生和大语言模型创建的头颈部超声选择题测试评估
Diagnostics (Basel). 2025 Jul 22;15(15):1848. doi: 10.3390/diagnostics15151848.
8
Advancements in artificial intelligence transforming medical education: a comprehensive overview.人工智能在医学教育中的进展:全面概述
Med Educ Online. 2025 Dec;30(1):2542807. doi: 10.1080/10872981.2025.2542807. Epub 2025 Aug 12.
9
Short- and long-term weekly patient-reported outcomes prediction undergoing radiotherapy: single-patient time series model vs. transformer-based multi-patient time series model.接受放射治疗患者的短期和长期每周自我报告结局预测:单患者时间序列模型与基于Transformer的多患者时间序列模型
BioData Min. 2025 Aug 12;18(1):53. doi: 10.1186/s13040-025-00464-7.
10
Ethical considerations and robustness of artificial neural networks in medical image analysis under data corruption.数据损坏情况下医学图像分析中人工神经网络的伦理考量与稳健性
Sci Rep. 2025 Aug 11;15(1):29305. doi: 10.1038/s41598-025-15268-2.

本文引用的文献

1
Evaluation and mitigation of cognitive biases in medical language models.医学语言模型中认知偏差的评估与缓解
NPJ Digit Med. 2024 Oct 21;7(1):295. doi: 10.1038/s41746-024-01283-6.
2
Zero-shot Learning with Minimum Instruction to Extract Social Determinants and Family History from Clinical Notes using GPT Model.使用GPT模型从临床记录中提取社会决定因素和家族病史的最少指令零样本学习
Proc IEEE Int Conf Big Data. 2023 Dec;2023:1476-1480. doi: 10.1109/BigData59044.2023.10386811.
3
Bias in artificial intelligence for medical imaging: fundamentals, detection, avoidance, mitigation, challenges, ethics, and prospects.医学成像人工智能中的偏差:基础、检测、避免、缓解、挑战、伦理及前景
Diagn Interv Radiol. 2025 Mar 3;31(2):75-88. doi: 10.4274/dir.2024.242854. Epub 2024 Jul 2.
4
The potential for artificial intelligence to transform healthcare: perspectives from international health leaders.人工智能改变医疗保健的潜力:国际卫生领导人的观点。
NPJ Digit Med. 2024 Apr 9;7(1):88. doi: 10.1038/s41746-024-01097-6.
5
Drug target prediction through deep learning functional representation of gene signatures.通过基因特征的深度学习功能表示进行药物靶标预测。
Nat Commun. 2024 Feb 29;15(1):1853. doi: 10.1038/s41467-024-46089-y.
6
Challenges and barriers of using large language models (LLM) such as ChatGPT for diagnostic medicine with a focus on digital pathology - a recent scoping review.使用大型语言模型(如 ChatGPT)进行诊断医学的挑战和障碍,重点是数字病理学——近期的范围综述。
Diagn Pathol. 2024 Feb 27;19(1):43. doi: 10.1186/s13000-024-01464-7.
7
Integrating Artificial Intelligence for Drug Discovery in the Context of Revolutionizing Drug Delivery.在药物递送变革的背景下整合人工智能用于药物发现。
Life (Basel). 2024 Feb 7;14(2):233. doi: 10.3390/life14020233.
8
Regression-based Deep-Learning predicts molecular biomarkers from pathology slides.基于回归的深度学习从病理切片中预测分子生物标志物。
Nat Commun. 2024 Feb 10;15(1):1253. doi: 10.1038/s41467-024-45589-1.
9
Improving health literacy using the power of digital communications to achieve better health outcomes for patients and practitioners.利用数字通信的力量提高健康素养,为患者和从业者带来更好的健康结果。
Front Digit Health. 2023 Nov 17;5:1264780. doi: 10.3389/fdgth.2023.1264780. eCollection 2023.
10
Impact of real-life use of artificial intelligence as support for human reading in a population-based breast cancer screening program with mammography and tomosynthesis.基于人群的乳腺癌筛查计划中,使用人工智能作为人类阅读支持的实际应用对乳腺 X 线摄影和断层合成的影响。
Eur Radiol. 2024 Jun;34(6):3958-3966. doi: 10.1007/s00330-023-10426-4. Epub 2023 Nov 17.