• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

Knowledge-enhanced Parameter-efficient Transfer Learning with METER for medical vision-language tasks.

作者信息

Liang Xudong, Xie Jiang, Wei Jinzhu, Zhang Mengfei, Zhang Haoyang

机构信息

School of Computer Engineering and Science, Shanghai University, Shanghai, China.

School of Computer Engineering and Science, Shanghai University, Shanghai, China.

出版信息

J Biomed Inform. 2025 Jun;166:104840. doi: 10.1016/j.jbi.2025.104840. Epub 2025 May 8.

DOI:10.1016/j.jbi.2025.104840
PMID:40348310
Abstract

OBJECTIVE

The full fine-tuning paradigm becomes impractical when applying pre-trained models to downstream tasks due to significant computational and storage costs. Parameter-efficient fine-tuning (PEFT) methods can alleviate the issue. However, solely applying PEFT methods leads to sub-optimal performance owing to the domain gap between pre-trained models and medical downstream tasks.

METHODS

This study proposes Knowledge-enhanced Parameter-efficient Transfer Learning with METER (KPL-METER) for medical vision-language (VL) downstream tasks. KPL-METER combines PEFT methods, including an innovative PEFT module for multi-modal branches and newly introduced external domain-specific knowledge to enhance model performance. First, a lightweight, plug-and-play module named Sharing Adapter (SAdapter) is developed and inserted into the multi-modal encoders. This allows the two modalities to maintain uni-modal features while encouraging cross-modal consistency. Second, a novel knowledge extraction method and a parameter-free knowledge modeling strategy are developed to incorporate domain-specific knowledge from the Unified Medical Language System (UMLS) into multi-modal features. To further enhance the modeling of uni-modal features, Adapter is added to the image and text encoders.

RESULTS

The effectiveness of the proposed model is evaluated on two medical VL tasks using three VL datasets. The results indicate that the KPL-METER model outperforms other PEFT methods in terms of performance while utilizing fewer parameters. Furthermore, KPL-METER-MED, which incorporates medical-tailored encoders, is developed. Compared to previous models in the medical domain, KPL-METER-MED tunes fewer parameters while generally achieving higher performance.

CONCLUSION

The proposed KPL-METER architecture effectively adapts general VL models for medical VL tasks, and the designed knowledge extraction and fusion method notably enhance performance by integrating medical domain-specific knowledge. Code is available at https://github.com/Adam-lxd/KPL-METER.

摘要

相似文献

1
Knowledge-enhanced Parameter-efficient Transfer Learning with METER for medical vision-language tasks.
J Biomed Inform. 2025 Jun;166:104840. doi: 10.1016/j.jbi.2025.104840. Epub 2025 May 8.
2
MCPL: Multi-Modal Collaborative Prompt Learning for Medical Vision-Language Model.MCPL:用于医学视觉语言模型的多模态协作提示学习
IEEE Trans Med Imaging. 2024 Dec;43(12):4224-4235. doi: 10.1109/TMI.2024.3418408. Epub 2024 Dec 2.
3
DVPT: Dynamic Visual Prompt Tuning of large pre-trained models for medical image analysis.DVPT:用于医学图像分析的大型预训练模型的动态视觉提示调整
Neural Netw. 2025 May;185:107168. doi: 10.1016/j.neunet.2025.107168. Epub 2025 Jan 16.
4
Parameter Efficient Fine-tuning of Transformer-based Masked Autoencoder Enhances Resource Constrained Neuroimage Analysis.基于Transformer的掩码自动编码器的参数高效微调增强了资源受限的神经图像分析。
bioRxiv. 2025 Feb 20:2025.02.15.638442. doi: 10.1101/2025.02.15.638442.
5
Efficiency at scale: Investigating the performance of diminutive language models in clinical tasks.规模化效率:研究微型语言模型在临床任务中的性能。
Artif Intell Med. 2024 Nov;157:103002. doi: 10.1016/j.artmed.2024.103002. Epub 2024 Oct 23.
6
Embedded prompt tuning: Towards enhanced calibration of pretrained models for medical images.嵌入式提示调整:增强医学图像预训练模型校准的新途径。
Med Image Anal. 2024 Oct;97:103258. doi: 10.1016/j.media.2024.103258. Epub 2024 Jul 4.
7
Cross-Modal self-supervised vision language pre-training with multiple objectives for medical visual question answering.用于医学视觉问答的多目标跨模态自监督视觉语言预训练
J Biomed Inform. 2024 Dec;160:104748. doi: 10.1016/j.jbi.2024.104748. Epub 2024 Nov 12.
8
AdaptFRCNet: Semi-supervised adaptation of pre-trained model with frequency and region consistency for medical image segmentation.AdaptFRCNet:基于频率和区域一致性的预训练模型半监督自适应医学图像分割方法
Med Image Anal. 2025 Jul;103:103626. doi: 10.1016/j.media.2025.103626. Epub 2025 May 13.
9
Democratizing Protein Language Models with Parameter-Efficient Fine-Tuning.通过参数高效微调实现蛋白质语言模型的民主化
bioRxiv. 2023 Nov 10:2023.11.09.566187. doi: 10.1101/2023.11.09.566187.
10
Democratizing protein language models with parameter-efficient fine-tuning.参数高效微调:用民主化方法对蛋白质语言模型进行优化。
Proc Natl Acad Sci U S A. 2024 Jun 25;121(26):e2405840121. doi: 10.1073/pnas.2405840121. Epub 2024 Jun 20.