• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

健康提示:一种临床自然语言处理的零样本学习范式。

HealthPrompt: A Zero-shot Learning Paradigm for Clinical Natural Language Processing.

机构信息

Intelligent Systems Program, School of Computing and Information, University of Pittsburgh, PA.

Department of Health Information Management, University of Pittsburgh, PA.

出版信息

AMIA Annu Symp Proc. 2023 Apr 29;2022:972-981. eCollection 2022.

PMID:37128372
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10148337/
Abstract

Developing clinical natural language systems based on machine learning and deep learning is dependent on the availability of large-scale annotated clinical text datasets, most of which are time-consuming to create and not publicly available. The lack of such annotated datasets is the biggest bottleneck for the development of clinical NLP systems. Zero-Shot Learning (ZSL) refers to the use of deep learning models to classify instances from new classes of which no training data have been seen before. Prompt-based learning is an emerging ZSL technique in NLP where we define task-based templates for different tasks. In this study, we developed a novel prompt-based clinical NLP framework called HealthPrompt and applied the paradigm of prompt-based learning on clinical texts. In this technique, rather than fine-tuning a Pre-trained Language Model (PLM), the task definitions are tuned by defining a prompt template. We performed an in-depth analysis of HealthPrompt on six different PLMs in a no-training-data setting. Our experiments show that HealthPrompt could effectively capture the context of clinical texts and perform well for clinical NLP tasks without any training data.

摘要

基于机器学习和深度学习开发临床自然语言系统依赖于大规模标注的临床文本数据集的可用性,而这些数据集大多需要耗费大量时间来创建,且无法公开获取。缺乏此类标注数据集是临床自然语言处理系统发展的最大瓶颈。零样本学习(ZSL)是指使用深度学习模型对以前从未见过训练数据的新类别的实例进行分类。基于提示的学习是 NLP 中一种新兴的 ZSL 技术,我们为不同任务定义基于任务的模板。在这项研究中,我们开发了一种名为 HealthPrompt 的新型基于提示的临床自然语言处理框架,并将基于提示的学习范式应用于临床文本。在这项技术中,不是通过微调预训练语言模型(PLM),而是通过定义提示模板来调整任务定义。我们在没有训练数据的情况下,在六个不同的 PLM 上对 HealthPrompt 进行了深入分析。我们的实验表明,HealthPrompt 可以有效地捕获临床文本的上下文,并在没有任何训练数据的情况下很好地执行临床自然语言处理任务。

相似文献

1
HealthPrompt: A Zero-shot Learning Paradigm for Clinical Natural Language Processing.健康提示:一种临床自然语言处理的零样本学习范式。
AMIA Annu Symp Proc. 2023 Apr 29;2022:972-981. eCollection 2022.
2
Few-Shot Learning for Clinical Natural Language Processing Using Siamese Neural Networks: Algorithm Development and Validation Study.使用暹罗神经网络的临床自然语言处理少样本学习:算法开发与验证研究
JMIR AI. 2023 May 4;2:e44293. doi: 10.2196/44293.
3
An Empirical Evaluation of Prompting Strategies for Large Language Models in Zero-Shot Clinical Natural Language Processing: Algorithm Development and Validation Study.零样本临床自然语言处理中大型语言模型提示策略的实证评估:算法开发与验证研究
JMIR Med Inform. 2024 Apr 8;12:e55318. doi: 10.2196/55318.
4
A clinical text classification paradigm using weak supervision and deep representation.一种使用弱监督和深度表示的临床文本分类范式。
BMC Med Inform Decis Mak. 2019 Jan 7;19(1):1. doi: 10.1186/s12911-018-0723-6.
5
Few-shot learning for medical text: A review of advances, trends, and opportunities.医学文本的少样本学习:进展、趋势和机遇综述。
J Biomed Inform. 2023 Aug;144:104458. doi: 10.1016/j.jbi.2023.104458. Epub 2023 Jul 23.
6
Annotated dataset creation through large language models for non-english medical NLP.通过大型语言模型创建非英语医学自然语言处理的标注数据集。
J Biomed Inform. 2023 Sep;145:104478. doi: 10.1016/j.jbi.2023.104478. Epub 2023 Aug 23.
7
Large language models for biomedicine: foundations, opportunities, challenges, and best practices.大型语言模型在生物医学领域的应用:基础、机遇、挑战和最佳实践。
J Am Med Inform Assoc. 2024 Sep 1;31(9):2114-2124. doi: 10.1093/jamia/ocae074.
8
Generative large language models are all-purpose text analytics engines: text-to-text learning is all your need.生成式大语言模型是通用文本分析引擎:文本到文本学习就是你所需要的一切。
J Am Med Inform Assoc. 2024 Sep 1;31(9):1892-1903. doi: 10.1093/jamia/ocae078.
9
AlpaPICO: Extraction of PICO frames from clinical trial documents using LLMs.AlpaPICO:使用大语言模型从临床试验文档中提取 PICO 框架。
Methods. 2024 Jun;226:78-88. doi: 10.1016/j.ymeth.2024.04.005. Epub 2024 Apr 21.
10
KEBLM: Knowledge-Enhanced Biomedical Language Models.KEBLM:知识增强型生物医学语言模型。
J Biomed Inform. 2023 Jul;143:104392. doi: 10.1016/j.jbi.2023.104392. Epub 2023 May 19.

引用本文的文献

1
Prompt Engineering in Clinical Practice: Tutorial for Clinicians.临床实践中的提示工程:临床医生教程
J Med Internet Res. 2025 Sep 15;27:e72644. doi: 10.2196/72644.
2
Keyword-optimized template insertion for clinical note classification via prompt-based learning.通过基于提示的学习进行关键词优化模板插入以实现临床笔记分类
BMC Med Inform Decis Mak. 2025 Jul 3;25(1):247. doi: 10.1186/s12911-025-03071-y.
3
Out of distribution learning in bioinformatics: advancements and challenges.生物信息学中的分布外学习:进展与挑战
Brief Bioinform. 2025 May 1;26(3). doi: 10.1093/bib/bbaf294.
4
When investigator meets large language models: a qualitative analysis of cancer patient decision-making journeys.当研究者遇上大语言模型:癌症患者决策历程的定性分析
NPJ Digit Med. 2025 Jun 5;8(1):336. doi: 10.1038/s41746-025-01747-3.
5
LLMs-based Few-Shot Disease Predictions using EHR: A Novel Approach Combining Predictive Agent Reasoning and Critical Agent Instruction.基于大语言模型的利用电子健康记录进行少样本疾病预测:一种结合预测性智能体推理与批判性智能体指令的新方法。
AMIA Annu Symp Proc. 2025 May 22;2024:319-328. eCollection 2024.
6
Year 2023 in Biomedical Natural Language Processing: a Tribute to Large Language Models and Generative AI.2023年生物医学自然语言处理领域:向大语言模型和生成式人工智能致敬。
Yearb Med Inform. 2024 Aug;33(1):241-248. doi: 10.1055/s-0044-1800751. Epub 2025 Apr 8.
7
Improving Dietary Supplement Information Retrieval: Development of a Retrieval-Augmented Generation System With Large Language Models.改善膳食补充剂信息检索:利用大语言模型开发检索增强生成系统
J Med Internet Res. 2025 Mar 19;27:e67677. doi: 10.2196/67677.
8
Information Extraction from Clinical Texts with Generative Pre-trained Transformer Models.利用生成式预训练Transformer模型从临床文本中提取信息。
Int J Med Sci. 2025 Feb 3;22(5):1015-1028. doi: 10.7150/ijms.103332. eCollection 2025.
9
Decoding substance use disorder severity from clinical notes using a large language model.使用大语言模型从临床记录中解码物质使用障碍的严重程度
Npj Ment Health Res. 2025 Feb 7;4(1):5. doi: 10.1038/s44184-024-00114-6.
10
CPLLM: Clinical prediction with large language models.CPLLM:基于大语言模型的临床预测
PLOS Digit Health. 2024 Dec 6;3(12):e0000680. doi: 10.1371/journal.pdig.0000680. eCollection 2024 Dec.

本文引用的文献

1
COVID-Twitter-BERT: A natural language processing model to analyse COVID-19 content on Twitter.COVID-Twitter-BERT:一种用于分析推特上新冠疫情相关内容的自然语言处理模型。
Front Artif Intell. 2023 Mar 14;6:1023281. doi: 10.3389/frai.2023.1023281. eCollection 2023.
2
Cohort selection for clinical trials: n2c2 2018 shared task track 1.队列选择用于临床试验:n2c2 2018 共享任务赛道 1。
J Am Med Inform Assoc. 2019 Nov 1;26(11):1163-1171. doi: 10.1093/jamia/ocz163.
3
BioBERT: a pre-trained biomedical language representation model for biomedical text mining.BioBERT:一种用于生物医学文本挖掘的预训练生物医学语言表示模型。
Bioinformatics. 2020 Feb 15;36(4):1234-1240. doi: 10.1093/bioinformatics/btz682.
4
Zero-Shot Learning-A Comprehensive Evaluation of the Good, the Bad and the Ugly.零样本学习:好坏丑的全面评估。
IEEE Trans Pattern Anal Mach Intell. 2019 Sep;41(9):2251-2265. doi: 10.1109/TPAMI.2018.2857768. Epub 2018 Jul 19.
5
MIMIC-III, a freely accessible critical care database.MIMIC-III,一个免费获取的重症监护数据库。
Sci Data. 2016 May 24;3:160035. doi: 10.1038/sdata.2016.35.
6
2010 i2b2/VA challenge on concepts, assertions, and relations in clinical text.2010 i2b2/VA 挑战赛:临床文本中的概念、断言和关系
J Am Med Inform Assoc. 2011 Sep-Oct;18(5):552-6. doi: 10.1136/amiajnl-2011-000203. Epub 2011 Jun 16.
7
Discovering peripheral arterial disease cases from radiology notes using natural language processing.使用自然语言处理技术从放射学记录中发现外周动脉疾病病例。
AMIA Annu Symp Proc. 2010 Nov 13;2010:722-6.
8
Active computerized pharmacovigilance using natural language processing, statistics, and electronic health records: a feasibility study.利用自然语言处理、统计学和电子健康记录的主动计算机化药物警戒:一项可行性研究。
J Am Med Inform Assoc. 2009 May-Jun;16(3):328-37. doi: 10.1197/jamia.M3028. Epub 2009 Mar 4.