Suppr超能文献

基于提示的自回归生成式多标签少样本ICD编码

Multi-label Few-shot ICD Coding as Autoregressive Generation with Prompt.

作者信息

Yang Zhichao, Kwon Sunjae, Yao Zonghai, Yu Hong

机构信息

College of Information and Computer Sciences, University of Massachusetts Amherst.

Department of Computer Science, University of Massachusetts Lowell.

出版信息

Proc AAAI Conf Artif Intell. 2023 Jun 26;37(4):5366-5374. doi: 10.1609/aaai.v37i4.25668.

Abstract

Automatic International Classification of Diseases (ICD) coding aims to assign multiple ICD codes to a medical note with an average of 3,000+ tokens. This task is challenging due to the high-dimensional space of multi-label assignment (155,000+ ICD code candidates) and the long-tail challenge - Many ICD codes are infrequently assigned yet infrequent ICD codes are important clinically. This study addresses the long-tail challenge by transforming this multi-label classification task into an autoregressive generation task. Specifically, we first introduce a novel pretraining objective to generate free text diagnoses and procedures using the SOAP structure, the medical logic physicians use for note documentation. Second, instead of directly predicting the high dimensional space of ICD codes, our model generates the lower dimension of text descriptions, which then infers ICD codes. Third, we designed a novel prompt template for multi-label classification. We evaluate our Generation with Prompt (GP) model with the benchmark of all code assignment (MIMIC-III-full) and few shot ICD code assignment evaluation benchmark (MIMIC-III-few). Experiments on MIMIC-III-few show that our model performs with a marco F130.2, which substantially outperforms the previous MIMIC-III-full SOTA model (marco F1 4.3) and the model specifically designed for few/zero shot setting (marco F1 18.7). Finally, we design a novel ensemble learner, a cross-attention reranker with prompts, to integrate previous SOTA and our best few-shot coding predictions. Experiments on MIMIC-III-full show that our ensemble learner substantially improves both macro and micro F1, from 10.4 to 14.6 and from 58.2 to 59.1, respectively.

摘要

自动国际疾病分类(ICD)编码旨在为平均包含3000多个词元的医学记录分配多个ICD编码。由于多标签分配的高维空间(超过155,000个ICD编码候选)以及长尾挑战,这项任务具有挑战性——许多ICD编码很少被分配,但罕见的ICD编码在临床上很重要。本研究通过将此多标签分类任务转化为自回归生成任务来应对长尾挑战。具体而言,我们首先引入一种新颖的预训练目标,使用SOAP结构(医生用于记录的医学逻辑)生成自由文本诊断和程序。其次,我们的模型不是直接预测ICD编码的高维空间,而是生成文本描述的低维表示,然后据此推断ICD编码。第三,我们设计了一种新颖的多标签分类提示模板。我们使用所有编码分配基准(MIMIC-III-full)和少样本ICD编码分配评估基准(MIMIC-III-few)对我们的带提示生成(GP)模型进行评估。在MIMIC-III-few上的实验表明,我们的模型的宏F1为30.2,大大优于之前的MIMIC-III-full最优模型(宏F1为4.3)以及专门为少样本/零样本设置设计的模型(宏F1为18.7)。最后,我们设计了一种新颖的集成学习器,即带提示的交叉注意力重排器,以整合之前的最优模型和我们最佳的少样本编码预测。在MIMIC-III-full上的实验表明,我们的集成学习器显著提高了宏F1和微F1,分别从10.4提高到14.6,从58.2提高到59.1。

相似文献

1
4
A Pseudo Label-Wise Attention Network for Automatic ICD Coding.基于伪标签注意力网络的 ICD 自动编码方法。
IEEE J Biomed Health Inform. 2022 Oct;26(10):5201-5212. doi: 10.1109/JBHI.2022.3193291. Epub 2022 Oct 5.
10
Can GPT-3.5 generate and code discharge summaries?GPT-3.5 可以生成和编写出院小结吗?
J Am Med Inform Assoc. 2024 Oct 1;31(10):2284-2293. doi: 10.1093/jamia/ocae132.

引用本文的文献

本文引用的文献

2
Clinical Prompt Learning With Frozen Language Models.临床提示学习与冻结语言模型。
IEEE Trans Neural Netw Learn Syst. 2024 Nov;35(11):16453-16463. doi: 10.1109/TNNLS.2023.3294633. Epub 2024 Oct 29.
7
Generating Accurate Electronic Health Assessment from Medical Graph.从医学图谱生成准确的电子健康评估。
Proc Conf Empir Methods Nat Lang Process. 2020 Nov;2020:3764-3773. doi: 10.18653/v1/2020.findings-emnlp.336.
9
Interpretable deep learning to map diagnostic texts to ICD-10 codes.可解释的深度学习将诊断文本映射到 ICD-10 代码。
Int J Med Inform. 2019 Sep;129:49-59. doi: 10.1016/j.ijmedinf.2019.05.015. Epub 2019 May 22.

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验