• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

OptimCLM:通过知识蒸馏、剪枝和量化优化临床语言模型以预测患者预后。

OptimCLM: Optimizing clinical language models for predicting patient outcomes via knowledge distillation, pruning and quantization.

作者信息

Hasan Mohammad Junayed, Rahman Fuad, Mohammed Nabeel

机构信息

Apurba NSU R&D Lab, Department of Electrical and Computer Engineering, North South University, Dhaka, Bangladesh.

Apurba Technologies Ltd., Dhaka, Bangladesh.

出版信息

Int J Med Inform. 2025 Mar;195:105764. doi: 10.1016/j.ijmedinf.2024.105764. Epub 2024 Dec 18.

DOI:10.1016/j.ijmedinf.2024.105764
PMID:39708669
Abstract

BACKGROUND

Clinical Language Models (CLMs) possess the potential to reform traditional healthcare systems by aiding in clinical decision making and optimal resource utilization. They can enhance patient outcomes and help healthcare management through predictive clinical tasks. However, their real-world deployment is limited due to high computational cost at inference, in terms of both time and space complexity.

OBJECTIVE

This study aims to develop and optimize an efficient framework that compresses CLMs without significant performance loss, reducing inference time and disk-space, and enabling real-world clinical applications.

METHODS

We introduce OptimCLM, a framework for optimizing CLMs with ensemble learning, knowledge distillation (KD), pruning and quantization. Based on domain-knowledge and performance, we select and combine domain-adaptive CLMs DischargeBERT and COReBERT as the teacher ensemble model. We transfer the teacher's knowledge to two smaller generalist models, BERT-PKD and TinyBERT, and apply black-box KD, post-training unstructured pruning and post-training 8-bit model quantization to them. In an admission-to-discharge setting, we evaluate the framework on four clinical outcome prediction tasks (length of stay prediction, mortality prediction, diagnosis prediction and procedure prediction) using admission notes from the MIMIC-III clinical database.

RESULTS

The OptimCLM framework achieved up to 22.88× compression ratio and 28.7× inference speedup, with less than 5% and 2% loss in macro-averaged AUROC for TinyBERT and BERT-PKD, respectively. The teacher model outperformed five state-of-the-art models on all tasks. The optimized BERT-PKD model also outperformed them in most tasks.

CONCLUSION

Our findings suggest that domain-specific fine-tuning with ensemble learning and KD is more effective than domain-specific pre-training for domain-knowledge transfer and text classification tasks. Thus, this work demonstrates the feasibility and potential of deploying optimized CLMs in healthcare settings and developing them with less computational resources.

摘要

背景

临床语言模型(CLMs)有潜力通过辅助临床决策和优化资源利用来改革传统医疗系统。它们可以改善患者预后,并通过预测性临床任务帮助医疗管理。然而,由于推理时在时间和空间复杂度方面的高计算成本,其在现实世界中的部署受到限制。

目的

本研究旨在开发和优化一个高效框架,该框架能在不显著损失性能的情况下压缩CLMs,减少推理时间和磁盘空间,并实现现实世界的临床应用。

方法

我们引入了OptimCLM,这是一个通过集成学习、知识蒸馏(KD)、剪枝和量化来优化CLMs的框架。基于领域知识和性能,我们选择并组合领域自适应CLMs DischargeBERT和COReBERT作为教师集成模型。我们将教师的知识转移到两个较小的通用模型BERT-PKD和TinyBERT,并对它们应用黑箱KD、训练后非结构化剪枝和训练后8位模型量化。在入院到出院的场景中,我们使用MIMIC-III临床数据库中的入院记录,在四个临床结局预测任务(住院时间预测、死亡率预测、诊断预测和手术预测)上评估该框架。

结果

OptimCLM框架实现了高达22.88倍的压缩率和28.7倍的推理加速,TinyBERT和BERT-PKD的宏平均AUROC损失分别小于5%和2%。教师模型在所有任务上均优于五个最先进的模型。优化后的BERT-PKD模型在大多数任务上也优于它们。

结论

我们的研究结果表明,对于领域知识转移和文本分类任务,使用集成学习和KD进行特定领域的微调比特定领域的预训练更有效。因此,这项工作证明了在医疗环境中部署优化后的CLMs并以更少的计算资源开发它们的可行性和潜力。

相似文献

1
OptimCLM: Optimizing clinical language models for predicting patient outcomes via knowledge distillation, pruning and quantization.OptimCLM:通过知识蒸馏、剪枝和量化优化临床语言模型以预测患者预后。
Int J Med Inform. 2025 Mar;195:105764. doi: 10.1016/j.ijmedinf.2024.105764. Epub 2024 Dec 18.
2
Distilling the knowledge from large-language model for health event prediction.从大语言模型中提取知识用于健康事件预测。
Sci Rep. 2024 Dec 28;14(1):30675. doi: 10.1038/s41598-024-75331-2.
3
SensiMix: Sensitivity-Aware 8-bit index & 1-bit value mixed precision quantization for BERT compression.SensiMix:用于 BERT 压缩的敏感度感知 8 位索引和 1 位值混合精度量化。
PLoS One. 2022 Apr 18;17(4):e0265621. doi: 10.1371/journal.pone.0265621. eCollection 2022.
4
DDK: Dynamic structure pruning based on differentiable search and recursive knowledge distillation for BERT.基于可微分搜索和递归知识蒸馏的 BERT 动态结构剪枝。
Neural Netw. 2024 May;173:106164. doi: 10.1016/j.neunet.2024.106164. Epub 2024 Feb 9.
5
Classifying social determinants of health from unstructured electronic health records using deep learning-based natural language processing.利用基于深度学习的自然语言处理技术从非结构化电子健康记录中分类社会健康决定因素。
J Biomed Inform. 2022 Mar;127:103984. doi: 10.1016/j.jbi.2021.103984. Epub 2022 Jan 7.
6
Autonomous International Classification of Diseases Coding Using Pretrained Language Models and Advanced Prompt Learning Techniques: Evaluation of an Automated Analysis System Using Medical Text.使用预训练语言模型和先进提示学习技术的自主国际疾病分类编码:对一个使用医学文本的自动分析系统的评估
JMIR Med Inform. 2025 Jan 6;13:e63020. doi: 10.2196/63020.
7
Multifaceted Natural Language Processing Task-Based Evaluation of Bidirectional Encoder Representations From Transformers Models for Bilingual (Korean and English) Clinical Notes: Algorithm Development and Validation.基于转换器模型的双向编码器表示的多方面自然语言处理任务评估在双语(韩语和英语)临床笔记中的应用:算法开发和验证。
JMIR Med Inform. 2024 Oct 30;12:e52897. doi: 10.2196/52897.
8
When BERT meets Bilbo: a learning curve analysis of pretrained language model on disease classification.当 BERT 遇见比尔博:预训练语言模型在疾病分类上的学习曲线分析。
BMC Med Inform Decis Mak. 2022 Apr 5;21(Suppl 9):377. doi: 10.1186/s12911-022-01829-2.
9
Extracting comprehensive clinical information for breast cancer using deep learning methods.利用深度学习方法提取乳腺癌全面临床信息。
Int J Med Inform. 2019 Dec;132:103985. doi: 10.1016/j.ijmedinf.2019.103985. Epub 2019 Oct 2.
10
Study of Deep Learning-Based Legal Judgment Prediction in Internet of Things Era.基于深度学习的物联网时代法律判决预测研究。
Comput Intell Neurosci. 2022 Aug 8;2022:8490760. doi: 10.1155/2022/8490760. eCollection 2022.

引用本文的文献

1
Artificial intelligence in electroencephalography analysis for epilepsy diagnosis and management.用于癫痫诊断和管理的脑电图分析中的人工智能
Front Neurol. 2025 Aug 18;16:1615120. doi: 10.3389/fneur.2025.1615120. eCollection 2025.
2
Early detection of occupational stress: Enhancing workplace safety with machine learning and large language models.职业压力的早期检测:利用机器学习和大语言模型提升工作场所安全性
PLoS One. 2025 Jun 2;20(6):e0323265. doi: 10.1371/journal.pone.0323265. eCollection 2025.