• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

利用大语言模型进行模拟心理治疗客户互动:Client101的开发与可用性研究

Leveraging Large Language Models for Simulated Psychotherapy Client Interactions: Development and Usability Study of Client101.

作者信息

Cabrera Lozoya Daniel, Conway Mike, Sebastiano De Duro Edoardo, D'Alfonso Simon

机构信息

School of Computing and Information Systems, The University of Melbourne, Gratham St, Parkville VIC, Melbourne, 3010, Australia, 61 90355511.

Department of Psychology and Cognitive Science, University of Trento, Trento, Italy.

出版信息

JMIR Med Educ. 2025 Jul 31;11:e68056. doi: 10.2196/68056.

DOI:10.2196/68056
PMID:40743466
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12312989/
Abstract

BACKGROUND

In recent years, large language models (LLMs) have shown a remarkable ability to generate human-like text. One potential application of this capability is using LLMs to simulate clients in a mental health context. This research presents the development and evaluation of Client101, a web conversational platform featuring LLM-driven chatbots designed to simulate mental health clients.

OBJECTIVE

We aim to develop and test a web-based conversational psychotherapy training tool designed to closely resemble clients with mental health issues.

METHODS

We used GPT-4 and prompt engineering techniques to develop chatbots that simulate realistic client conversations. Two chatbots were created based on clinical vignette cases: one representing a person with depression and the other, a person with generalized anxiety disorder. A total of 16 mental health professionals were instructed to conduct single sessions with the chatbots using a cognitive behavioral therapy framework; a total of 15 sessions with the anxiety chatbot and 14 with the depression chatbot were completed. After each session, participants completed a 19-question survey assessing the chatbot's ability to simulate the mental health condition and its potential as a training tool. Additionally, we used the LIWC (Linguistic Inquiry and Word Count) tool to analyze the psycholinguistic features of the chatbot conversations related to anxiety and depression. These features were compared to those in a set of webchat psychotherapy sessions with human clients-42 sessions related to anxiety and 47 related to depression-using an independent samples t test.

RESULTS

Participants' survey responses were predominantly positive regarding the chatbots' realism and portrayal of mental health conditions. For instance, 93% (14/15) considered that the chatbot provided a coherent and convincing narrative typical of someone with an anxiety condition. The statistical analysis of LIWC psycholinguistic features revealed significant differences between chatbot and human therapy transcripts for 3 of 8 anxiety-related features: negations (t56=4.03, P=.001), family (t56=-8.62, P=.001), and negative emotions (t56=-3.91, P=.002). The remaining 5 features-sadness, personal pronouns, present focus, social, and anger-did not show significant differences. For depression-related features, 4 of 9 showed significant differences: negative emotions (t60=-3.84, P=.003), feeling (t60=-6.40, P<.001), health (t60=-4.13, P=.001), and illness (t60=-5.52, P<.001). The other 5 features-sadness, anxiety, mental, first-person pronouns, and discrepancy-did not show statistically significant differences.

CONCLUSIONS

This research underscores both the strengths and limitations of using GPT-4-powered chatbots as tools for psychotherapy training. Participant feedback suggests that the chatbots effectively portray mental health conditions and are generally perceived as valuable training aids. However, differences in specific psycholinguistic features suggest targeted areas for enhancement, helping refine Client101's effectiveness as a tool for training mental health professionals.

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c2bb/12312989/ee7a85d25d10/mededu-v11-e68056-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c2bb/12312989/9ea43c7d3a1a/mededu-v11-e68056-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c2bb/12312989/4f2df0e04f3f/mededu-v11-e68056-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c2bb/12312989/bf8810fe1a0f/mededu-v11-e68056-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c2bb/12312989/9e64d4a10704/mededu-v11-e68056-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c2bb/12312989/a976d83a21f0/mededu-v11-e68056-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c2bb/12312989/ee7a85d25d10/mededu-v11-e68056-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c2bb/12312989/9ea43c7d3a1a/mededu-v11-e68056-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c2bb/12312989/4f2df0e04f3f/mededu-v11-e68056-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c2bb/12312989/bf8810fe1a0f/mededu-v11-e68056-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c2bb/12312989/9e64d4a10704/mededu-v11-e68056-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c2bb/12312989/a976d83a21f0/mededu-v11-e68056-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c2bb/12312989/ee7a85d25d10/mededu-v11-e68056-g006.jpg
摘要

背景

近年来,大语言模型(LLMs)展现出了生成类人文本的卓越能力。这种能力的一个潜在应用是在心理健康环境中使用大语言模型来模拟来访者。本研究介绍了Client101的开发与评估,这是一个基于网络的对话平台,其特色是由大语言模型驱动的聊天机器人,旨在模拟心理健康来访者。

目的

我们旨在开发并测试一个基于网络的对话式心理治疗训练工具,该工具旨在逼真地模拟有心理健康问题的来访者。

方法

我们使用GPT-4和提示工程技术来开发模拟真实来访者对话的聊天机器人。基于临床病例创建了两个聊天机器人:一个代表抑郁症患者,另一个代表广泛性焦虑症患者。共有16名心理健康专业人员被要求使用认知行为疗法框架与聊天机器人进行单次会话;与焦虑症聊天机器人共完成了15次会话,与抑郁症聊天机器人共完成了14次会话。每次会话后,参与者完成一项包含19个问题的调查,评估聊天机器人模拟心理健康状况的能力及其作为训练工具的潜力。此外,我们使用语言查询与字数统计(LIWC)工具来分析与焦虑和抑郁相关的聊天机器人对话的心理语言学特征。使用独立样本t检验将这些特征与一组与人类来访者进行的网络聊天心理治疗会话中的特征进行比较,其中与焦虑相关的会话有42次,与抑郁相关的会话有47次。

结果

参与者对聊天机器人的真实感和对心理健康状况的描绘的调查反馈大多是积极的。例如,93%(14/15)的人认为聊天机器人提供了符合焦虑症患者典型特征的连贯且有说服力的叙述。对LIWC心理语言学特征的统计分析显示,在8个与焦虑相关的特征中,有3个在聊天机器人和人类治疗记录之间存在显著差异:否定词(t56 = 4.03,P = 0.001)、家庭相关词汇(t56 = -8.62,P = 0.001)和负面情绪(t56 = -3.91,P = 0.002)。其余5个特征——悲伤、人称代词、当前关注、社交和愤怒——没有显示出显著差异。对于与抑郁相关的特征,9个中有4个显示出显著差异:负面情绪(t60 = -3.84,P = 0.003)、情感(t60 = -6.40,P < 0.001)、健康(t60 = -4.13,P = 0.001)和疾病(t60 = -5.52,P < 0.001)。其他5个特征——悲伤、焦虑、心理、第一人称代词和差异——没有显示出统计学上的显著差异。

结论

本研究强调了使用由GPT-4驱动的聊天机器人作为心理治疗训练工具的优势和局限性。参与者的反馈表明,聊天机器人有效地描绘了心理健康状况,并且通常被视为有价值的训练辅助工具。然而,特定心理语言学特征的差异表明了需要改进的目标领域,有助于提高Client101作为训练心理健康专业人员工具的有效性。

相似文献

1
Leveraging Large Language Models for Simulated Psychotherapy Client Interactions: Development and Usability Study of Client101.利用大语言模型进行模拟心理治疗客户互动:Client101的开发与可用性研究
JMIR Med Educ. 2025 Jul 31;11:e68056. doi: 10.2196/68056.
2
Chatbot for the Return of Positive Genetic Screening Results for Hereditary Cancer Syndromes: Prompt Engineering Project.遗传性癌症综合征阳性基因筛查结果返回的聊天机器人:提示工程设计项目
JMIR Cancer. 2025 Jun 10;11:e65848. doi: 10.2196/65848.
3
Future Me, a Prospection-Based Chatbot to Promote Mental Well-Being in Youth: Two Exploratory User Experience Studies.“未来的我”,一款基于展望的聊天机器人,用于促进青少年的心理健康:两项探索性用户体验研究。
JMIR Form Res. 2025 Jul 18;9:e74411. doi: 10.2196/74411.
4
Development and Validation of a Large Language Model-Powered Chatbot for Neurosurgery: Mixed Methods Study on Enhancing Perioperative Patient Education.用于神经外科手术的基于大语言模型的聊天机器人的开发与验证:关于加强围手术期患者教育的混合方法研究
J Med Internet Res. 2025 Jul 15;27:e74299. doi: 10.2196/74299.
5
Development of a GPT-4-Powered Virtual Simulated Patient and Communication Training Platform for Medical Students to Practice Discussing Abnormal Mammogram Results With Patients: Multiphase Study.开发一个由GPT-4驱动的虚拟模拟患者和沟通训练平台,供医学生练习与患者讨论异常乳房X光检查结果:多阶段研究。
JMIR Form Res. 2025 Apr 17;9:e65670. doi: 10.2196/65670.
6
Psychological therapies for treatment-resistant depression in adults.成人难治性抑郁症的心理治疗
Cochrane Database Syst Rev. 2018 May 14;5(5):CD010558. doi: 10.1002/14651858.CD010558.pub2.
7
The Black Book of Psychotropic Dosing and Monitoring.《精神药物剂量与监测黑皮书》
Psychopharmacol Bull. 2024 Jul 8;54(3):8-59.
8
Investigation and analysis of mental health status of the older adult in western rural areas.西部农村地区老年人心理健康状况的调查与分析
Front Public Health. 2025 Jul 16;13:1612600. doi: 10.3389/fpubh.2025.1612600. eCollection 2025.
9
Client feedback in psychological therapy for children and adolescents with mental health problems.针对有心理健康问题的儿童和青少年的心理治疗中的客户反馈。
Cochrane Database Syst Rev. 2018 Aug 20;8(8):CD011729. doi: 10.1002/14651858.CD011729.pub2.
10
Sexual Harassment and Prevention Training性骚扰与预防培训

本文引用的文献

1
Exploring the Ethical Challenges of Conversational AI in Mental Health Care: Scoping Review.探索心理健康护理中对话式人工智能的伦理挑战:范围审查
JMIR Ment Health. 2025 Feb 21;12:e60432. doi: 10.2196/60432.
2
Toward Linguistic Recognition of Generalized Anxiety Disorder.迈向广泛性焦虑症的语言识别
Front Digit Health. 2022 Apr 15;4:779039. doi: 10.3389/fdgth.2022.779039. eCollection 2022.
3
Perceptions and Opinions of Patients About Mental Health Chatbots: Scoping Review.患者对心理健康聊天机器人的看法和意见:范围综述。
J Med Internet Res. 2021 Jan 13;23(1):e17828. doi: 10.2196/17828.
4
Males and Mental Health Stigma.男性与心理健康污名化。
Am J Mens Health. 2020 Jul-Aug;14(4):1557988320949322. doi: 10.1177/1557988320949322.
5
A Novel Multimodal Digital Service (Moderated Online Social Therapy+) for Help-Seeking Young People Experiencing Mental Ill-Health: Pilot Evaluation Within a National Youth E-Mental Health Service.一种新颖的多模态数字服务(在线社交治疗+),用于帮助有心理健康问题的年轻寻求帮助者:国家青年电子心理健康服务中的试点评估。
J Med Internet Res. 2020 Aug 13;22(8):e17155. doi: 10.2196/17155.
6
Methods in predictive techniques for mental health status on social media: a critical review.社交媒体上心理健康状况预测技术的方法:批判性综述
NPJ Digit Med. 2020 Mar 24;3:43. doi: 10.1038/s41746-020-0233-7. eCollection 2020.
7
Development and Evaluation of ClientBot: Patient-Like Conversational Agent to Train Basic Counseling Skills.ClientBot的开发与评估:用于培训基本咨询技能的类患者对话代理
J Med Internet Res. 2019 Jul 15;21(7):e12529. doi: 10.2196/12529.
8
Chatbots and Conversational Agents in Mental Health: A Review of the Psychiatric Landscape.聊天机器人和会话代理在心理健康中的应用:精神病学领域的综述。
Can J Psychiatry. 2019 Jul;64(7):456-464. doi: 10.1177/0706743719828977. Epub 2019 Mar 21.
9
Facebook language predicts depression in medical records.脸书的语言可预测病历中的抑郁症状。
Proc Natl Acad Sci U S A. 2018 Oct 30;115(44):11203-11208. doi: 10.1073/pnas.1802331115. Epub 2018 Oct 15.
10
Boundary crossings and violations in clinical settings.临床环境中的越界与违规行为。
Indian J Psychol Med. 2012 Jan;34(1):21-4. doi: 10.4103/0253-7176.96151.