• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于临床自然语言处理的轻量级变压器

Lightweight transformers for clinical natural language processing.

作者信息

Rohanian Omid, Nouriborji Mohammadmahdi, Jauncey Hannah, Kouchaki Samaneh, Nooralahzadeh Farhad, Clifton Lei, Merson Laura, Clifton David A

机构信息

Department of Engineering Science, University of Oxford, Oxford, UK.

NLPie Research, Oxford, UK.

出版信息

Nat Lang Eng. 2024 Sep;30(5):887-914. doi: 10.1017/S1351324923000542. Epub 2024 Jan 12.

DOI:10.1017/S1351324923000542
PMID:39588066
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11586117/
Abstract

Specialised pre-trained language models are becoming more frequent in Natural language Processing (NLP) since they can potentially outperform models trained on generic texts. BioBERT (Sanh et al., Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter. , 2019) and BioClinicalBERT (Alsentzer et al., Publicly available clinical bert embeddings. In , pp. 72-78, 2019) are two examples of such models that have shown promise in medical NLP tasks. Many of these models are overparametrised and resource-intensive, but thanks to techniques like knowledge distillation, it is possible to create smaller versions that perform almost as well as their larger counterparts. In this work, we specifically focus on development of compact language models for processing clinical texts (i.e. progress notes, discharge summaries, etc). We developed a number of efficient lightweight clinical transformers using knowledge distillation and continual learning, with the number of parameters ranging from million to million. These models performed comparably to larger models such as BioBERT and ClinicalBioBERT and significantly outperformed other compact models trained on general or biomedical data. Our extensive evaluation was done across several standard datasets and covered a wide range of clinical text-mining tasks, including natural language inference, relation extraction, named entity recognition and sequence classification. To our knowledge, this is the first comprehensive study specifically focused on creating efficient and compact transformers for clinical NLP tasks. The models and code used in this study can be found on our Huggingface profile at https://huggingface.co/nlpie and Github page at https://github.com/nlpie-research/Lightweight-Clinical-Transformers, respectively, promoting reproducibility of our results.

摘要

专门的预训练语言模型在自然语言处理(NLP)中越来越常见,因为它们有可能优于基于通用文本训练的模型。BioBERT(桑赫等人,Distilbert,BERT的精简版:更小、更快、更便宜、更轻便。,2019年)和BioClinicalBERT(阿尔森策等人,公开可用的临床BERT嵌入。在,第72 - 78页,2019年)就是这类模型的两个例子,它们在医学NLP任务中显示出了潜力。这些模型中的许多都参数过多且资源密集,但由于知识蒸馏等技术,可以创建几乎与较大版本性能相同的较小版本。在这项工作中,我们特别专注于开发用于处理临床文本(即病程记录、出院小结等)的紧凑语言模型。我们使用知识蒸馏和持续学习开发了一些高效的轻量级临床变压器,参数数量从百万到百万不等。这些模型的表现与BioBERT和ClinicalBioBERT等较大模型相当,并且明显优于在通用或生物医学数据上训练的其他紧凑模型。我们在几个标准数据集上进行了广泛评估,涵盖了广泛的临床文本挖掘任务,包括自然语言推理、关系提取、命名实体识别和序列分类。据我们所知,这是第一项专门专注于为临床NLP任务创建高效紧凑变压器的综合研究。本研究中使用的模型和代码分别可以在我们的Huggingface页面https://huggingface.co/nlpie和Github页面https://github.com/nlpie-research/Lightweight-Clinical-Transformers上找到,这有助于我们结果的可重复性。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b6b4/11586117/91601fc90371/S1351324923000542_fig4.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b6b4/11586117/b124330a45f3/S1351324923000542_fig1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b6b4/11586117/9100ea54c7a6/S1351324923000542_fig2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b6b4/11586117/7a2827e4611e/S1351324923000542_fig3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b6b4/11586117/91601fc90371/S1351324923000542_fig4.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b6b4/11586117/b124330a45f3/S1351324923000542_fig1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b6b4/11586117/9100ea54c7a6/S1351324923000542_fig2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b6b4/11586117/7a2827e4611e/S1351324923000542_fig3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b6b4/11586117/91601fc90371/S1351324923000542_fig4.jpg

相似文献

1
Lightweight transformers for clinical natural language processing.用于临床自然语言处理的轻量级变压器
Nat Lang Eng. 2024 Sep;30(5):887-914. doi: 10.1017/S1351324923000542. Epub 2024 Jan 12.
2
On the effectiveness of compact biomedical transformers.紧凑型生物医学变压器的有效性。
Bioinformatics. 2023 Mar 1;39(3). doi: 10.1093/bioinformatics/btad103.
3
BioBERT: a pre-trained biomedical language representation model for biomedical text mining.BioBERT:一种用于生物医学文本挖掘的预训练生物医学语言表示模型。
Bioinformatics. 2020 Feb 15;36(4):1234-1240. doi: 10.1093/bioinformatics/btz682.
4
Bioformer: an efficient transformer language model for biomedical text mining.生物former:一种用于生物医学文本挖掘的高效Transformer语言模型。
ArXiv. 2023 Feb 3:arXiv:2302.01588v1.
5
Few-Shot Learning for Clinical Natural Language Processing Using Siamese Neural Networks: Algorithm Development and Validation Study.使用暹罗神经网络的临床自然语言处理少样本学习:算法开发与验证研究
JMIR AI. 2023 May 4;2:e44293. doi: 10.2196/44293.
6
Multifaceted Natural Language Processing Task-Based Evaluation of Bidirectional Encoder Representations From Transformers Models for Bilingual (Korean and English) Clinical Notes: Algorithm Development and Validation.基于转换器模型的双向编码器表示的多方面自然语言处理任务评估在双语(韩语和英语)临床笔记中的应用:算法开发和验证。
JMIR Med Inform. 2024 Oct 30;12:e52897. doi: 10.2196/52897.
7
A comparative study of pretrained language models for long clinical text.基于预训练语言模型的长临床文本比较研究
J Am Med Inform Assoc. 2023 Jan 18;30(2):340-347. doi: 10.1093/jamia/ocac225.
8
BioBERT and Similar Approaches for Relation Extraction.BioBERT 及其在关系抽取中的应用。
Methods Mol Biol. 2022;2496:221-235. doi: 10.1007/978-1-0716-2305-3_12.
9
Annotated dataset creation through large language models for non-english medical NLP.通过大型语言模型创建非英语医学自然语言处理的标注数据集。
J Biomed Inform. 2023 Sep;145:104478. doi: 10.1016/j.jbi.2023.104478. Epub 2023 Aug 23.
10
Extracting comprehensive clinical information for breast cancer using deep learning methods.利用深度学习方法提取乳腺癌全面临床信息。
Int J Med Inform. 2019 Dec;132:103985. doi: 10.1016/j.ijmedinf.2019.103985. Epub 2019 Oct 2.

引用本文的文献

1
Exploring the effectiveness of instruction tuning in biomedical language processing.探索指令调优在生物医学语言处理中的有效性。
Artif Intell Med. 2024 Dec;158:103007. doi: 10.1016/j.artmed.2024.103007. Epub 2024 Nov 7.
2
The future of multimodal artificial intelligence models for integrating imaging and clinical metadata: a narrative review.整合影像学与临床元数据的多模态人工智能模型的未来:一篇综述
Diagn Interv Radiol. 2024 Oct 1. doi: 10.4274/dir.2024.242631.

本文引用的文献

1
On the effectiveness of compact biomedical transformers.紧凑型生物医学变压器的有效性。
Bioinformatics. 2023 Mar 1;39(3). doi: 10.1093/bioinformatics/btad103.
2
Clinical utility of automatic phenotype annotation in unstructured clinical notes: intensive care unit use.临床实用:无结构临床记录中的自动表型标注——重症监护病房的应用
BMJ Health Care Inform. 2022 Nov;29(1). doi: 10.1136/bmjhci-2021-100519.
3
ISARIC-COVID-19 dataset: A Prospective, Standardized, Global Dataset of Patients Hospitalized with COVID-19.ISARIC-COVID-19 数据集:一项前瞻性、标准化、全球 COVID-19 住院患者数据集。
Sci Data. 2022 Jul 30;9(1):454. doi: 10.1038/s41597-022-01534-9.
4
Clinical notes: An untapped opportunity for improving risk prediction for hospitalization and emergency department visit during home health care.临床记录:改善居家医疗期间住院和急诊就诊风险预测的一个未开发机会。
J Biomed Inform. 2022 Apr;128:104039. doi: 10.1016/j.jbi.2022.104039. Epub 2022 Feb 26.
5
Unstructured clinical notes within the 24 hours since admission predict short, mid & long-term mortality in adult ICU patients.入院后 24 小时内的非结构化临床记录可预测成人 ICU 患者的短期、中期和长期死亡率。
PLoS One. 2022 Jan 6;17(1):e0262182. doi: 10.1371/journal.pone.0262182. eCollection 2022.
6
A weakly supervised model for the automated detection of adverse events using clinical notes.一种使用临床记录自动检测不良事件的弱监督模型。
J Biomed Inform. 2022 Feb;126:103969. doi: 10.1016/j.jbi.2021.103969. Epub 2021 Dec 3.
7
Identification of asthma control factor in clinical notes using a hybrid deep learning model.使用混合深度学习模型从临床记录中识别哮喘控制因素。
BMC Med Inform Decis Mak. 2021 Nov 9;21(Suppl 7):272. doi: 10.1186/s12911-021-01633-4.
8
The value of open-source clinical science in pandemic response: lessons from ISARIC.开源临床科学在应对大流行中的价值:来自国际严重急性呼吸道感染和新发呼吸道病毒合作组织(ISARIC)的经验教训。
Lancet Infect Dis. 2021 Dec;21(12):1623-1624. doi: 10.1016/S1473-3099(21)00565-X. Epub 2021 Oct 4.
9
Extracting Adverse Drug Events from Clinical Notes.从临床记录中提取药物不良反应。
AMIA Jt Summits Transl Sci Proc. 2021 May 17;2021:420-429. eCollection 2021.
10
Natural Language Processing of Clinical Notes to Identify Mental Illness and Substance Use Among People Living with HIV: Retrospective Cohort Study.利用临床记录的自然语言处理技术识别HIV感染者中的精神疾病和药物使用情况:回顾性队列研究
JMIR Med Inform. 2021 Mar 10;9(3):e23456. doi: 10.2196/23456.