• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

利用大型语言模型的零样本和少样本学习能力进行监管研究。

Harnessing large language models' zero-shot and few-shot learning capabilities for regulatory research.

机构信息

Division of Applied Regulatory Science, Office of Clinical Pharmacology, Office of Translational Sciences, Center for Drug Evaluation and Research, U.S. Food and Drug Administration, WO Bldg 64, 10903 New Hampshire Ave, Silver Spring, MD 20993, United States.

出版信息

Brief Bioinform. 2024 Jul 25;25(5). doi: 10.1093/bib/bbae354.

DOI:10.1093/bib/bbae354
PMID:39177261
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11342240/
Abstract

Large language models (LLMs) are sophisticated AI-driven models trained on vast sources of natural language data. They are adept at generating responses that closely mimic human conversational patterns. One of the most notable examples is OpenAI's ChatGPT, which has been extensively used across diverse sectors. Despite their flexibility, a significant challenge arises as most users must transmit their data to the servers of companies operating these models. Utilizing ChatGPT or similar models online may inadvertently expose sensitive information to the risk of data breaches. Therefore, implementing LLMs that are open source and smaller in scale within a secure local network becomes a crucial step for organizations where ensuring data privacy and protection has the highest priority, such as regulatory agencies. As a feasibility evaluation, we implemented a series of open-source LLMs within a regulatory agency's local network and assessed their performance on specific tasks involving extracting relevant clinical pharmacology information from regulatory drug labels. Our research shows that some models work well in the context of few- or zero-shot learning, achieving performance comparable, or even better than, neural network models that needed thousands of training samples. One of the models was selected to address a real-world issue of finding intrinsic factors that affect drugs' clinical exposure without any training or fine-tuning. In a dataset of over 700 000 sentences, the model showed a 78.5% accuracy rate. Our work pointed to the possibility of implementing open-source LLMs within a secure local network and using these models to perform various natural language processing tasks when large numbers of training examples are unavailable.

摘要

大型语言模型(LLMs)是一种基于大量自然语言数据训练而成的复杂 AI 驱动模型。它们擅长生成与人类对话模式非常相似的回复。OpenAI 的 ChatGPT 是最著名的例子之一,已在多个领域得到广泛应用。尽管它们具有灵活性,但由于大多数用户必须将数据传输到运营这些模型的公司服务器,因此面临一个重大挑战。在线使用 ChatGPT 或类似模型可能会无意中将敏感信息暴露在数据泄露的风险中。因此,对于那些将数据隐私和保护视为首要任务的组织(如监管机构)来说,实施开源的、规模较小的 LLM,并将其置于安全的本地网络中,成为至关重要的一步。作为可行性评估,我们在一个监管机构的本地网络中实现了一系列开源 LLM,并评估了它们在特定任务上的性能,这些任务涉及从监管药物标签中提取相关临床药理学信息。我们的研究表明,在少量或零样本学习的情况下,某些模型的性能良好,甚至可以与需要数千个训练样本的神经网络模型相媲美。我们选择了其中一个模型来解决一个实际问题,即在没有任何训练或微调的情况下,找出影响药物临床暴露的内在因素。在一个超过 70 万条句子的数据集上,该模型的准确率达到了 78.5%。我们的工作表明,在安全的本地网络中实施开源 LLM 并使用这些模型在无法获得大量训练示例的情况下执行各种自然语言处理任务是可能的。

相似文献

1
Harnessing large language models' zero-shot and few-shot learning capabilities for regulatory research.利用大型语言模型的零样本和少样本学习能力进行监管研究。
Brief Bioinform. 2024 Jul 25;25(5). doi: 10.1093/bib/bbae354.
2
Prescription of Controlled Substances: Benefits and Risks管制药品的处方:益处与风险
3
Short-Term Memory Impairment短期记忆障碍
4
Sexual Harassment and Prevention Training性骚扰与预防培训
5
The first step is the hardest: pitfalls of representing and tokenizing temporal data for large language models.第一步是最困难的:为大型语言模型表示和标记时间数据的陷阱。
J Am Med Inform Assoc. 2024 Sep 1;31(9):2151-2158. doi: 10.1093/jamia/ocae090.
6
Stench of Errors or the Shine of Potential: The Challenge of (Ir)Responsible Use of ChatGPT in Speech-Language Pathology.错误的恶臭还是潜力的光辉:言语病理学中(不)负责任地使用ChatGPT的挑战。
Int J Lang Commun Disord. 2025 Jul-Aug;60(4):e70088. doi: 10.1111/1460-6984.70088.
7
Evaluating the Reasoning Capabilities of Large Language Models for Medical Coding and Hospital Readmission Risk Stratification: Zero-Shot Prompting Approach.评估大型语言模型在医学编码和医院再入院风险分层方面的推理能力:零样本提示方法。
J Med Internet Res. 2025 Jul 30;27:e74142. doi: 10.2196/74142.
8
A dataset and benchmark for hospital course summarization with adapted large language models.一个用于医院病程总结的数据集和基准测试,采用了适配的大语言模型。
J Am Med Inform Assoc. 2025 Mar 1;32(3):470-479. doi: 10.1093/jamia/ocae312.
9
Implementing Large Language Models in Health Care: Clinician-Focused Review With Interactive Guideline.在医疗保健中应用大语言模型:以临床医生为重点的回顾与交互式指南
J Med Internet Res. 2025 Jul 11;27:e71916. doi: 10.2196/71916.
10
The Black Book of Psychotropic Dosing and Monitoring.《精神药物剂量与监测黑皮书》
Psychopharmacol Bull. 2024 Jul 8;54(3):8-59.

引用本文的文献

1
Should LLMs be over empowered for high-stake regulatory research?对于高风险的监管研究,大型语言模型是否被赋予了过多权力?
Brief Bioinform. 2025 May 1;26(3). doi: 10.1093/bib/bbaf299.

本文引用的文献

1
Ensemble pretrained language models to extract biomedical knowledge from literature.基于预训练语言模型的方法从文献中提取生物医学知识。
J Am Med Inform Assoc. 2024 Sep 1;31(9):1904-1911. doi: 10.1093/jamia/ocae061.
2
Deep learning-enabled natural language processing to identify directional pharmacokinetic drug-drug interactions.深度学习赋能的自然语言处理用于识别有方向的药代动力学药物相互作用。
BMC Bioinformatics. 2023 Nov 1;24(1):413. doi: 10.1186/s12859-023-05520-9.
3
A dataset for plain language adaptation of biomedical abstracts.生物医学文摘的自然语言适应数据集。
Sci Data. 2023 Jan 4;10(1):8. doi: 10.1038/s41597-022-01920-3.
4
A general procedure to select calibration drugs for lab-specific validation and calibration of proarrhythmia risk prediction models: An illustrative example using the CiPA model.一种用于选择校准药物的通用程序,用于实验室特定验证和心律失常风险预测模型的校准:使用 CiPA 模型的说明性示例。
J Pharmacol Toxicol Methods. 2020 Sep;105:106890. doi: 10.1016/j.vascn.2020.106890. Epub 2020 Jun 21.
5
A systematic strategy for estimating hERG block potency and its implications in a new cardiac safety paradigm.一种用于评估 hERG 阻断效力的系统策略及其在新的心脏安全性范式中的意义。
Toxicol Appl Pharmacol. 2020 May 1;394:114961. doi: 10.1016/j.taap.2020.114961. Epub 2020 Mar 21.
6
Fine-Tuning Bidirectional Encoder Representations From Transformers (BERT)-Based Models on Large-Scale Electronic Health Record Notes: An Empirical Study.基于大规模电子健康记录笔记对基于变换器的双向编码器表征(BERT)模型进行微调:一项实证研究。
JMIR Med Inform. 2019 Sep 12;7(3):e14830. doi: 10.2196/14830.
7
BioBERT: a pre-trained biomedical language representation model for biomedical text mining.BioBERT:一种用于生物医学文本挖掘的预训练生物医学语言表示模型。
Bioinformatics. 2020 Feb 15;36(4):1234-1240. doi: 10.1093/bioinformatics/btz682.
8
A guide to deep learning in healthcare.深度学习在医疗保健中的应用指南。
Nat Med. 2019 Jan;25(1):24-29. doi: 10.1038/s41591-018-0316-z. Epub 2019 Jan 7.
9
Assessment of an In Silico Mechanistic Model for Proarrhythmia Risk Prediction Under the CiPA Initiative.基于 CiPA 倡议的致心律失常风险预测的计算机制模型评估。
Clin Pharmacol Ther. 2019 Feb;105(2):466-475. doi: 10.1002/cpt.1184. Epub 2018 Aug 27.
10
A Unified approach for Conventional Zero-shot, Generalized Zero-shot and Few-shot Learning.一种用于传统零样本、广义零样本和少样本学习的统一方法。
IEEE Trans Image Process. 2018 Jul 31. doi: 10.1109/TIP.2018.2861573.