IEEE Trans Neural Netw Learn Syst. 2024 Nov;35(11):16453-16463. doi: 10.1109/TNNLS.2023.3294633. Epub 2024 Oct 29.
When the first transformer-based language models were published in the late 2010s, pretraining with general text and then fine-tuning the model on a task-specific dataset often achieved the state-of-the-art performance. However, more recent work suggests that for some tasks, directly prompting the pretrained model matches or surpasses fine-tuning in performance with few or no model parameter updates required. The use of prompts with language models for natural language processing (NLP) tasks is known as prompt learning. We investigated the viability of prompt learning on clinically meaningful decision tasks and directly compared this with more traditional fine-tuning methods. Results show that prompt learning methods were able to match or surpass the performance of traditional fine-tuning with up to 1000 times fewer trainable parameters, less training time, less training data, and lower computation resource requirements. We argue that these characteristics make prompt learning a very desirable alternative to traditional fine-tuning for clinical tasks, where the computational resources of public health providers are limited, and where data can often not be made available or not be used for fine-tuning due to patient privacy concerns. The complementary code to reproduce the experiments presented in this work can be found at https://github.com/NtaylorOX/Public_Clinical_Prompt.
当基于变压器的语言模型在 2010 年代后期首次发布时,使用通用文本进行预训练,然后在特定于任务的数据集上微调模型,通常可以实现最先进的性能。然而,最近的研究表明,对于某些任务,直接提示预训练模型的性能与微调相当,甚至在不需要更新模型参数的情况下超过微调。在自然语言处理 (NLP) 任务中使用提示与语言模型一起被称为提示学习。我们研究了提示学习在具有临床意义的决策任务中的可行性,并直接将其与更传统的微调方法进行了比较。结果表明,提示学习方法能够在使用多达 1000 倍更少的可训练参数、更少的训练时间、更少的训练数据和更低的计算资源要求的情况下,匹配或超过传统微调的性能。我们认为,这些特性使得提示学习成为传统微调的一种非常理想的替代方法,适用于公共卫生提供者的计算资源有限的临床任务,并且由于患者隐私问题,数据通常无法用于微调。可在 https://github.com/NtaylorOX/Public_Clinical_Prompt 找到重现本文中提出的实验的补充代码。