• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

神经提示搜索

Neural Prompt Search.

作者信息

Zhang Yuanhan, Zhou Kaiyang, Liu Ziwei

出版信息

IEEE Trans Pattern Anal Mach Intell. 2025 Jul;47(7):5268-5280. doi: 10.1109/TPAMI.2024.3435939.

DOI:10.1109/TPAMI.2024.3435939
PMID:39078759
Abstract

The size of vision models has grown exponentially over the last few years, especially after the emergence of Vision Transformer. This has motivated the development of parameter-efficient tuning methods, such as learning adapter layers or visual prompt tokens, which allow a tiny portion of model parameters to be trained whereas the vast majority obtained from pre-training are frozen. However, designing a proper tuning method is non-trivial: one might need to try out a lengthy list of design choices, not to mention that each downstream dataset often requires custom designs. In this paper, we view the existing parameter-efficient tuning methods as "prompt modules" and propose Neural prOmpt seArcH (NOAH), a novel approach that learns, for large vision models, the optimal design of prompt modules through a neural architecture search algorithm, specifically for each downstream dataset. By conducting extensive experiments on over 20 vision datasets, we demonstrate that NOAH (i) is superior to individual prompt modules, (ii) has good few-shot learning ability, and (iii) is domain-generalizable.

摘要

在过去几年中,视觉模型的规模呈指数级增长,尤其是在视觉Transformer出现之后。这推动了参数高效调整方法的发展,例如学习适配器层或视觉提示令牌,这些方法允许只训练一小部分模型参数,而从预训练中获得的绝大多数参数则保持冻结状态。然而,设计一种合适的调整方法并非易事:可能需要尝试一长串的设计选择,更不用说每个下游数据集通常都需要定制设计了。在本文中,我们将现有的参数高效调整方法视为“提示模块”,并提出了神经提示搜索(NOAH),这是一种新颖的方法,它通过神经架构搜索算法为大型视觉模型学习提示模块的最优设计,特别是针对每个下游数据集。通过在20多个视觉数据集上进行广泛的实验,我们证明了NOAH(i)优于单个提示模块,(ii)具有良好的少样本学习能力,以及(iii)具有领域通用性。

相似文献

1
Neural Prompt Search.神经提示搜索
IEEE Trans Pattern Anal Mach Intell. 2025 Jul;47(7):5268-5280. doi: 10.1109/TPAMI.2024.3435939.
2
Stitched vision transformer for age-related macular degeneration detection using retinal optical coherence tomography images.基于视网膜光学相干断层扫描图像的老年性黄斑变性检测用缝合视觉Transformer。
PLoS One. 2024 Jun 5;19(6):e0304943. doi: 10.1371/journal.pone.0304943. eCollection 2024.
3
Prompt tuning for parameter-efficient medical image segmentation.用于参数高效医学图像分割的提示调优
Med Image Anal. 2024 Jan;91:103024. doi: 10.1016/j.media.2023.103024. Epub 2023 Nov 7.
4
Embedded prompt tuning: Towards enhanced calibration of pretrained models for medical images.嵌入式提示调整:增强医学图像预训练模型校准的新途径。
Med Image Anal. 2024 Oct;97:103258. doi: 10.1016/j.media.2024.103258. Epub 2024 Jul 4.
5
DVPT: Dynamic Visual Prompt Tuning of large pre-trained models for medical image analysis.DVPT:用于医学图像分析的大型预训练模型的动态视觉提示调整
Neural Netw. 2025 May;185:107168. doi: 10.1016/j.neunet.2025.107168. Epub 2025 Jan 16.
6
Model tuning or prompt Tuning? a study of large language models for clinical concept and relation extraction.模型调优还是提示调优?大型语言模型在临床概念和关系抽取中的应用研究。
J Biomed Inform. 2024 May;153:104630. doi: 10.1016/j.jbi.2024.104630. Epub 2024 Mar 26.
7
Learning Domain Invariant Prompt for Vision-Language Models.用于视觉语言模型的学习领域不变提示
IEEE Trans Image Process. 2024;33:1348-1360. doi: 10.1109/TIP.2024.3362062. Epub 2024 Feb 14.
8
A prompt regularization approach to enhance few-shot class-incremental learning with Two-Stage Classifier.
Neural Netw. 2025 Aug;188:107453. doi: 10.1016/j.neunet.2025.107453. Epub 2025 Apr 7.
9
Class-Specific Prompt Learning for Vision-Language Models.
IEEE Trans Neural Netw Learn Syst. 2025 Sep;36(9):16213-16222. doi: 10.1109/TNNLS.2025.3566559.
10
Proto-Adapter: Efficient Training-Free CLIP-Adapter for Few-Shot Image Classification.Proto-Adapter:用于少样本图像分类的高效无需训练的CLIP-Adapter
Sensors (Basel). 2024 Jun 4;24(11):3624. doi: 10.3390/s24113624.

引用本文的文献

1
EIM: An effective solution for improving multi-modal large language models.EIM:一种改进多模态大语言模型的有效解决方案。
PLoS One. 2025 Aug 11;20(8):e0329590. doi: 10.1371/journal.pone.0329590. eCollection 2025.