Suppr超能文献

基于预训练语言模型的长临床文本比较研究

A comparative study of pretrained language models for long clinical text.

机构信息

Division of Health and Biomedical Informatics, Department of Preventive Medicine, Feinberg School of Medicine, Northwestern University, Chicago, Illinois, USA.

Division of Cardiology, Department of Medicine, Feinberg School of Medicine, Northwestern University, Chicago, Illinois, USA.

出版信息

J Am Med Inform Assoc. 2023 Jan 18;30(2):340-347. doi: 10.1093/jamia/ocac225.

Abstract

OBJECTIVE

Clinical knowledge-enriched transformer models (eg, ClinicalBERT) have state-of-the-art results on clinical natural language processing (NLP) tasks. One of the core limitations of these transformer models is the substantial memory consumption due to their full self-attention mechanism, which leads to the performance degradation in long clinical texts. To overcome this, we propose to leverage long-sequence transformer models (eg, Longformer and BigBird), which extend the maximum input sequence length from 512 to 4096, to enhance the ability to model long-term dependencies in long clinical texts.

MATERIALS AND METHODS

Inspired by the success of long-sequence transformer models and the fact that clinical notes are mostly long, we introduce 2 domain-enriched language models, Clinical-Longformer and Clinical-BigBird, which are pretrained on a large-scale clinical corpus. We evaluate both language models using 10 baseline tasks including named entity recognition, question answering, natural language inference, and document classification tasks.

RESULTS

The results demonstrate that Clinical-Longformer and Clinical-BigBird consistently and significantly outperform ClinicalBERT and other short-sequence transformers in all 10 downstream tasks and achieve new state-of-the-art results.

DISCUSSION

Our pretrained language models provide the bedrock for clinical NLP using long texts. We have made our source code available at https://github.com/luoyuanlab/Clinical-Longformer, and the pretrained models available for public download at: https://huggingface.co/yikuan8/Clinical-Longformer.

CONCLUSION

This study demonstrates that clinical knowledge-enriched long-sequence transformers are able to learn long-term dependencies in long clinical text. Our methods can also inspire the development of other domain-enriched long-sequence transformers.

摘要

目的

临床知识增强型转换器模型(例如 ClinicalBERT)在临床自然语言处理(NLP)任务中取得了最先进的成果。这些转换器模型的核心局限性之一是由于其全自注意力机制而导致的大量内存消耗,这导致在长临床文本中性能下降。为了克服这一问题,我们建议利用长序列转换器模型(例如 Longformer 和 BigBird),将最大输入序列长度从 512 扩展到 4096,以增强对长临床文本中长时依赖关系建模的能力。

材料和方法

受长序列转换器模型成功的启发,以及临床笔记大多较长的事实,我们引入了 2 个领域增强型语言模型 Clinical-Longformer 和 Clinical-BigBird,它们是在大规模临床语料库上进行预训练的。我们使用包括命名实体识别、问答、自然语言推理和文档分类任务在内的 10 个基准任务来评估这两种语言模型。

结果

结果表明,Clinical-Longformer 和 Clinical-BigBird 在所有 10 个下游任务中始终如一地显著优于 ClinicalBERT 和其他短序列转换器,并实现了新的最先进的结果。

讨论

我们的预训练语言模型为使用长文本进行临床 NLP 提供了基础。我们已经将源代码发布在 https://github.com/luoyuanlab/Clinical-Longformer 上,并在 https://huggingface.co/yikuan8/Clinical-Longformer 上提供了预训练模型供公众下载。

结论

本研究表明,临床知识增强型长序列转换器能够学习长临床文本中的长时依赖关系。我们的方法还可以启发其他领域增强型长序列转换器的发展。

相似文献

3
Clinical concept extraction using transformers.使用转换器进行临床概念提取。
J Am Med Inform Assoc. 2020 Dec 9;27(12):1935-1942. doi: 10.1093/jamia/ocaa189.
10
KEBLM: Knowledge-Enhanced Biomedical Language Models.KEBLM:知识增强型生物医学语言模型。
J Biomed Inform. 2023 Jul;143:104392. doi: 10.1016/j.jbi.2023.104392. Epub 2023 May 19.

引用本文的文献

5
Synthetic4Health: generating annotated synthetic clinical letters.合成4健康:生成带注释的合成临床信件。
Front Digit Health. 2025 May 30;7:1497130. doi: 10.3389/fdgth.2025.1497130. eCollection 2025.

本文引用的文献

3
Limitations of Transformers on Clinical Text Classification.Transformer 在临床文本分类上的局限性。
IEEE J Biomed Health Inform. 2021 Sep;25(9):3596-3607. doi: 10.1109/JBHI.2021.3062322. Epub 2021 Sep 3.
4
Early Prediction of Acute Kidney Injury in Critical Care Setting Using Clinical Notes.利用临床记录对重症监护环境中的急性肾损伤进行早期预测。
Proceedings (IEEE Int Conf Bioinformatics Biomed). 2018 Dec;2018:683-686. doi: 10.1109/bibm.2018.8621574. Epub 2019 Jan 24.

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验