• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

医学大语言模型容易受到数据中毒攻击。

Medical large language models are vulnerable to data-poisoning attacks.

作者信息

Alber Daniel Alexander, Yang Zihao, Alyakin Anton, Yang Eunice, Rai Sumedha, Valliani Aly A, Zhang Jeff, Rosenbaum Gabriel R, Amend-Thomas Ashley K, Kurland David B, Kremer Caroline M, Eremiev Alexander, Negash Bruck, Wiggan Daniel D, Nakatsuka Michelle A, Sangwon Karl L, Neifert Sean N, Khan Hammad A, Save Akshay Vinod, Palla Adhith, Grin Eric A, Hedman Monika, Nasir-Moin Mustafa, Liu Xujin Chris, Jiang Lavender Yao, Mankowski Michal A, Segev Dorry L, Aphinyanaphongs Yindalon, Riina Howard A, Golfinos John G, Orringer Daniel A, Kondziolka Douglas, Oermann Eric Karl

机构信息

Department of Neurosurgery, NYU Langone Health, New York, NY, USA.

New York University Grossman School of Medicine, New York, NY, USA.

出版信息

Nat Med. 2025 Feb;31(2):618-626. doi: 10.1038/s41591-024-03445-1. Epub 2025 Jan 8.

DOI:10.1038/s41591-024-03445-1
PMID:39779928
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11835729/
Abstract

The adoption of large language models (LLMs) in healthcare demands a careful analysis of their potential to spread false medical knowledge. Because LLMs ingest massive volumes of data from the open Internet during training, they are potentially exposed to unverified medical knowledge that may include deliberately planted misinformation. Here, we perform a threat assessment that simulates a data-poisoning attack against The Pile, a popular dataset used for LLM development. We find that replacement of just 0.001% of training tokens with medical misinformation results in harmful models more likely to propagate medical errors. Furthermore, we discover that corrupted models match the performance of their corruption-free counterparts on open-source benchmarks routinely used to evaluate medical LLMs. Using biomedical knowledge graphs to screen medical LLM outputs, we propose a harm mitigation strategy that captures 91.9% of harmful content (F1 = 85.7%). Our algorithm provides a unique method to validate stochastically generated LLM outputs against hard-coded relationships in knowledge graphs. In view of current calls for improved data provenance and transparent LLM development, we hope to raise awareness of emergent risks from LLMs trained indiscriminately on web-scraped data, particularly in healthcare where misinformation can potentially compromise patient safety.

摘要

在医疗保健领域采用大语言模型(LLMs)需要仔细分析其传播错误医学知识的可能性。由于大语言模型在训练期间会从开放互联网摄取大量数据,它们有可能接触到未经证实的医学知识,其中可能包括故意植入的错误信息。在此,我们进行了一项威胁评估,模拟针对用于大语言模型开发的流行数据集The Pile的数据中毒攻击。我们发现,仅用0.001%的医学错误信息替换训练令牌就会导致有害模型更有可能传播医疗错误。此外,我们发现,在常用于评估医学大语言模型的开源基准测试中,被破坏的模型与未被破坏的模型表现相当。通过使用生物医学知识图谱来筛选医学大语言模型的输出,我们提出了一种减轻危害的策略,该策略能够捕捉91.9%的有害内容(F1 = 85.7%)。我们的算法提供了一种独特的方法,可根据知识图谱中的硬编码关系验证随机生成的大语言模型输出。鉴于当前对改进数据来源和透明的大语言模型开发的呼吁,我们希望提高人们对在网络抓取数据上进行无差别训练的大语言模型所带来的新出现风险的认识,尤其是在医疗保健领域,错误信息可能会对患者安全造成潜在威胁。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be4/11835729/56dcc207919b/41591_2024_3445_Fig9_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be4/11835729/9f6f30561efc/41591_2024_3445_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be4/11835729/336385c74328/41591_2024_3445_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be4/11835729/5b817091861e/41591_2024_3445_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be4/11835729/4df055608a8b/41591_2024_3445_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be4/11835729/860a3cc9683c/41591_2024_3445_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be4/11835729/18d86c6c892f/41591_2024_3445_Fig6_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be4/11835729/4d2b1154ea2c/41591_2024_3445_Fig7_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be4/11835729/75c7a32dde66/41591_2024_3445_Fig8_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be4/11835729/56dcc207919b/41591_2024_3445_Fig9_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be4/11835729/9f6f30561efc/41591_2024_3445_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be4/11835729/336385c74328/41591_2024_3445_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be4/11835729/5b817091861e/41591_2024_3445_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be4/11835729/4df055608a8b/41591_2024_3445_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be4/11835729/860a3cc9683c/41591_2024_3445_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be4/11835729/18d86c6c892f/41591_2024_3445_Fig6_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be4/11835729/4d2b1154ea2c/41591_2024_3445_Fig7_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be4/11835729/75c7a32dde66/41591_2024_3445_Fig8_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be4/11835729/56dcc207919b/41591_2024_3445_Fig9_ESM.jpg

相似文献

1
Medical large language models are vulnerable to data-poisoning attacks.医学大语言模型容易受到数据中毒攻击。
Nat Med. 2025 Feb;31(2):618-626. doi: 10.1038/s41591-024-03445-1. Epub 2025 Jan 8.
2
Medical Misinformation in AI-Assisted Self-Diagnosis: Development of a Method (EvalPrompt) for Analyzing Large Language Models.人工智能辅助自我诊断中的医学错误信息:一种用于分析大语言模型的方法(EvalPrompt)的开发
JMIR Form Res. 2025 Mar 10;9:e66207. doi: 10.2196/66207.
3
Using Generative Artificial Intelligence in Health Economics and Outcomes Research: A Primer on Techniques and Breakthroughs.在卫生经济学与结果研究中使用生成式人工智能:技术与突破入门
Pharmacoecon Open. 2025 Apr 29. doi: 10.1007/s41669-025-00580-4.
4
A dataset and benchmark for hospital course summarization with adapted large language models.一个用于医院病程总结的数据集和基准测试,采用了适配的大语言模型。
J Am Med Inform Assoc. 2025 Mar 1;32(3):470-479. doi: 10.1093/jamia/ocae312.
5
An Automatic and End-to-End System for Rare Disease Knowledge Graph Construction Based on Ontology-Enhanced Large Language Models: Development Study.基于本体增强大语言模型的罕见病知识图谱构建自动端到端系统:开发研究
JMIR Med Inform. 2024 Dec 18;12:e60665. doi: 10.2196/60665.
6
Potential of Large Language Models in Health Care: Delphi Study.大语言模型在医疗保健中的潜力:德尔菲研究。
J Med Internet Res. 2024 May 13;26:e52399. doi: 10.2196/52399.
7
Large Language Models in Worldwide Medical Exams: Platform Development and Comprehensive Analysis.全球医学考试中的大语言模型:平台开发与综合分析
J Med Internet Res. 2024 Dec 27;26:e66114. doi: 10.2196/66114.
8
Large Language Models and User Trust: Consequence of Self-Referential Learning Loop and the Deskilling of Health Care Professionals.大语言模型与用户信任:自我参照学习循环的后果及医疗保健专业人员的技能退化
J Med Internet Res. 2024 Apr 25;26:e56764. doi: 10.2196/56764.
9
Identification of Online Health Information Using Large Pretrained Language Models: Mixed Methods Study.使用大型预训练语言模型识别在线健康信息:混合方法研究。
J Med Internet Res. 2025 May 14;27:e70733. doi: 10.2196/70733.
10
Exposing Vulnerabilities in Clinical LLMs Through Data Poisoning Attacks: Case Study in Breast Cancer.通过数据中毒攻击揭示临床语言模型中的漏洞:乳腺癌案例研究
medRxiv. 2024 Mar 21:2024.03.20.24304627. doi: 10.1101/2024.03.20.24304627.

引用本文的文献

1
AI Agents in Clinical Medicine: A Systematic Review.临床医学中的人工智能代理:一项系统综述。
medRxiv. 2025 Aug 26:2025.08.22.25334232. doi: 10.1101/2025.08.22.25334232.
2
Artificial intelligence in echocardiography: current applications and future perspectives.超声心动图中的人工智能:当前应用与未来展望。
J Echocardiogr. 2025 Aug 21. doi: 10.1007/s12574-025-00703-0.
3
Use of a Medical Communication Framework to Assess the Quality of Generative Artificial Intelligence Replies to Primary Care Patient Portal Messages: Content Analysis.

本文引用的文献

1
Medical large language models are susceptible to targeted misinformation attacks.医学大语言模型容易受到针对性错误信息攻击。
NPJ Digit Med. 2024 Oct 23;7(1):288. doi: 10.1038/s41746-024-01282-7.
2
Detecting hallucinations in large language models using semantic entropy.使用语义熵检测大型语言模型中的幻觉。
Nature. 2024 Jun;630(8017):625-630. doi: 10.1038/s41586-024-07421-0. Epub 2024 Jun 19.
3
MedCPT: Contrastive Pre-trained Transformers with large-scale PubMed search logs for zero-shot biomedical information retrieval.
使用医学交流框架评估生成式人工智能对基层医疗患者门户消息的回复质量:内容分析
JMIR Form Res. 2025 Jul 31;9:e71966. doi: 10.2196/71966.
4
DeepSeek-R1 outperforms Gemini 2.0 Pro, OpenAI o1, and o3-mini in bilingual complex ophthalmology reasoning.在双语复杂眼科推理方面,DeepSeek-R1的表现优于Gemini 2.0 Pro、OpenAI的o1和o3-mini。
Adv Ophthalmol Pract Res. 2025 May 9;5(3):189-195. doi: 10.1016/j.aopr.2025.05.001. eCollection 2025 Aug-Sep.
5
Implementing Large Language Models in Health Care: Clinician-Focused Review With Interactive Guideline.在医疗保健中应用大语言模型:以临床医生为重点的回顾与交互式指南
J Med Internet Res. 2025 Jul 11;27:e71916. doi: 10.2196/71916.
6
Clinical Management of Wasp Stings Using Large Language Models: Cross-Sectional Evaluation Study.使用大语言模型对黄蜂蜇伤进行临床管理:横断面评估研究
J Med Internet Res. 2025 Jun 4;27:e67489. doi: 10.2196/67489.
7
Large Language Models in Cancer Imaging: Applications and Future Perspectives.癌症成像中的大语言模型:应用与未来展望。
J Clin Med. 2025 May 8;14(10):3285. doi: 10.3390/jcm14103285.
8
The Applications of Large Language Models in Mental Health: Scoping Review.大语言模型在心理健康领域的应用:范围综述
J Med Internet Res. 2025 May 5;27:e69284. doi: 10.2196/69284.
9
Benchmark evaluation of DeepSeek large language models in clinical decision-making.临床决策中DeepSeek大语言模型的基准评估。
Nat Med. 2025 Apr 23. doi: 10.1038/s41591-025-03727-2.
10
Closing the Automation Gap: Robust AI for Dual-Stain Cervical Cancer Screening Triage.弥合自动化差距:用于双染宫颈癌筛查分诊的强大人工智能
Res Sq. 2025 Mar 4:rs.3.rs-5985837. doi: 10.21203/rs.3.rs-5985837/v1.
MedCPT:利用大规模 PubMed 检索日志进行零样本生物医学信息检索的对比预训练 Transformer。
Bioinformatics. 2023 Nov 1;39(11). doi: 10.1093/bioinformatics/btad651.
4
Large language models propagate race-based medicine.大语言模型传播基于种族的医学观念。
NPJ Digit Med. 2023 Oct 20;6(1):195. doi: 10.1038/s41746-023-00939-z.
5
Large language models in medicine.医学中的大型语言模型。
Nat Med. 2023 Aug;29(8):1930-1940. doi: 10.1038/s41591-023-02448-8. Epub 2023 Jul 17.
6
Large language models encode clinical knowledge.大语言模型编码临床知识。
Nature. 2023 Aug;620(7972):172-180. doi: 10.1038/s41586-023-06291-2. Epub 2023 Jul 12.
7
Health system-scale language models are all-purpose prediction engines.健康系统规模的语言模型是通用的预测引擎。
Nature. 2023 Jul;619(7969):357-362. doi: 10.1038/s41586-023-06160-y. Epub 2023 Jun 7.
8
BioGPT: generative pre-trained transformer for biomedical text generation and mining.BioGPT:用于生物医学文本生成和挖掘的生成式预训练转换器。
Brief Bioinform. 2022 Nov 19;23(6). doi: 10.1093/bib/bbac409.
9
Learning a Health Knowledge Graph from Electronic Medical Records.从电子病历中学习健康知识图谱。
Sci Rep. 2017 Jul 20;7(1):5994. doi: 10.1038/s41598-017-05778-z.
10
Systematic Poisoning Attacks on and Defenses for Machine Learning in Healthcare.系统中毒攻击与机器学习在医疗保健中的防御
IEEE J Biomed Health Inform. 2015 Nov;19(6):1893-905. doi: 10.1109/JBHI.2014.2344095. Epub 2014 Jul 30.