Lau Clinton, Zhu Xiaodan, Chan Wai-Yip
Department of Electrical and Computer Engineering & Ingenuity Labs, Queen's University, Kingston, ON, Canada.
Front Psychiatry. 2023 Jun 15;14:1160291. doi: 10.3389/fpsyt.2023.1160291. eCollection 2023.
To assist mental health care providers with the assessment of depression, research to develop a standardized, accessible, and non-invasive technique has garnered considerable attention. Our study focuses on the application of deep learning models for automatic assessment of depression severity based on clinical interview transcriptions. Despite the recent success of deep learning, the lack of large-scale high-quality datasets is a major performance bottleneck for many mental health applications.
A novel approach is proposed to address the data scarcity problem for depression assessment. It leverages both pretrained large language models and parameter-efficient tuning techniques. The approach is built upon adapting a small set of tunable parameters, known as prefix vectors, to guide a pretrained model towards predicting the Patient Health Questionnaire (PHQ)-8 score of a person. Experiments were conducted on the Distress Analysis Interview Corpus - Wizard of Oz (DAIC-WOZ) benchmark dataset with 189 subjects, partitioned into training, development, and test sets. Model learning was done on the training set. Prediction performance mean and standard deviation of each model, with five randomly-initialized runs, were reported on the development set. Finally, optimized models were evaluated on the test set.
The proposed model with prefix vectors outperformed all previously published methods, including models which utilized multiple types of data modalities, and achieved the best reported performance on the test set of DAIC-WOZ with a root mean square error of 4.67 and a mean absolute error of 3.80 on the PHQ-8 scale. Compared to conventionally fine-tuned baseline models, prefix-enhanced models were less prone to overfitting by using far fewer training parameters (<6% relatively).
While transfer learning through pretrained large language models can provide a good starting point for downstream learning, prefix vectors can further adapt the pretrained models effectively to the depression assessment task by only adjusting a small number of parameters. The improvement is in part due to the fine-grain flexibility of prefix vector size in adjusting the model's learning capacity. Our results provide evidence that prefix-tuning can be a useful approach in developing tools for automatic depression assessment.
为帮助心理健康护理人员评估抑郁症,开发一种标准化、可获取且非侵入性技术的研究已备受关注。我们的研究聚焦于基于临床访谈转录本,运用深度学习模型自动评估抑郁症严重程度。尽管深度学习近期取得了成功,但缺乏大规模高质量数据集仍是许多心理健康应用的主要性能瓶颈。
提出一种新颖方法来解决抑郁症评估的数据稀缺问题。它利用预训练的大语言模型和参数高效调整技术。该方法基于调整一小组可调参数(称为前缀向量)构建,以引导预训练模型预测一个人的患者健康问卷(PHQ)-8得分。在包含189名受试者的痛苦分析访谈语料库 - 绿野仙踪(DAIC-WOZ)基准数据集上进行实验,该数据集被划分为训练集、开发集和测试集。模型学习在训练集上进行。报告了每个模型在开发集上五次随机初始化运行的预测性能均值和标准差。最后,在测试集上评估优化后的模型。
所提出的带前缀向量的模型优于所有先前发表的方法,包括利用多种数据模态的模型,并在DAIC-WOZ测试集上取得了最佳报告性能,在PHQ-8量表上均方根误差为4.67,平均绝对误差为3.80。与传统微调的基线模型相比,前缀增强模型使用少得多的训练参数(相对少<6%),不太容易过拟合。
虽然通过预训练大语言模型进行迁移学习可为下游学习提供良好起点,但前缀向量仅通过调整少量参数就能使预训练模型有效适应抑郁症评估任务。这种改进部分归因于前缀向量大小在调整模型学习能力方面的细粒度灵活性。我们的结果证明前缀调整可能是开发自动抑郁症评估工具的一种有用方法。