Department of Psychiatry, New York University, Grossman School of Medicine, New York, NY, 10016, USA.
Talkspace, New York, NY, 10025, USA.
Transl Psychiatry. 2023 Oct 6;13(1):309. doi: 10.1038/s41398-023-02592-2.
Neuropsychiatric disorders pose a high societal cost, but their treatment is hindered by lack of objective outcomes and fidelity metrics. AI technologies and specifically Natural Language Processing (NLP) have emerged as tools to study mental health interventions (MHI) at the level of their constituent conversations. However, NLP's potential to address clinical and research challenges remains unclear. We therefore conducted a pre-registered systematic review of NLP-MHI studies using PRISMA guidelines (osf.io/s52jh) to evaluate their models, clinical applications, and to identify biases and gaps. Candidate studies (n = 19,756), including peer-reviewed AI conference manuscripts, were collected up to January 2023 through PubMed, PsycINFO, Scopus, Google Scholar, and ArXiv. A total of 102 articles were included to investigate their computational characteristics (NLP algorithms, audio features, machine learning pipelines, outcome metrics), clinical characteristics (clinical ground truths, study samples, clinical focus), and limitations. Results indicate a rapid growth of NLP MHI studies since 2019, characterized by increased sample sizes and use of large language models. Digital health platforms were the largest providers of MHI data. Ground truth for supervised learning models was based on clinician ratings (n = 31), patient self-report (n = 29) and annotations by raters (n = 26). Text-based features contributed more to model accuracy than audio markers. Patients' clinical presentation (n = 34), response to intervention (n = 11), intervention monitoring (n = 20), providers' characteristics (n = 12), relational dynamics (n = 14), and data preparation (n = 4) were commonly investigated clinical categories. Limitations of reviewed studies included lack of linguistic diversity, limited reproducibility, and population bias. A research framework is developed and validated (NLPxMHI) to assist computational and clinical researchers in addressing the remaining gaps in applying NLP to MHI, with the goal of improving clinical utility, data access, and fairness.
神经精神疾病造成了很高的社会成本,但由于缺乏客观的结果和保真度指标,其治疗受到了阻碍。人工智能技术,特别是自然语言处理(NLP)已成为研究心理健康干预(MHI)的工具,可以从其组成对话的层面进行研究。然而,NLP 解决临床和研究挑战的潜力仍不清楚。因此,我们按照 PRISMA 指南(osf.io/s52jh)对 NLP-MHI 研究进行了预先注册的系统评价,以评估其模型、临床应用,并确定偏见和差距。通过 PubMed、PsycINFO、Scopus、Google Scholar 和 ArXiv 收集了截止到 2023 年 1 月的候选研究(n=19756),包括同行评审的人工智能会议手稿。共有 102 篇文章被纳入研究,以调查其计算特征(NLP 算法、音频特征、机器学习管道、结果指标)、临床特征(临床真实情况、研究样本、临床重点)和局限性。结果表明,自 2019 年以来,NLP-MHI 研究呈快速增长趋势,其特点是样本量增加和大型语言模型的使用。数字健康平台是 MHI 数据的最大提供者。监督学习模型的真实数据基于临床医生的评分(n=31)、患者的自我报告(n=29)和评分者的注释(n=26)。基于文本的特征对模型准确性的贡献大于音频标记。研究中常见的临床类别包括患者的临床表现(n=34)、对干预的反应(n=11)、干预监测(n=20)、提供者的特征(n=12)、关系动态(n=14)和数据准备(n=4)。综述研究的局限性包括缺乏语言多样性、可重复性有限和人口偏差。开发并验证了一个研究框架(NLPxMHI),以帮助计算和临床研究人员解决在将 NLP 应用于 MHI 方面的剩余差距,目标是提高临床实用性、数据获取和公平性。