Franke Föyen Ludwig, Zapel Emma, Lekander Mats, Hedman-Lagerlöf Erik, Lindsäter Elin
Division of Psychology, Department of Clinical Neuroscience, Karolinska Institutet, Stockholm, Sweden.
Stress Research Institute, Department of Psychology, Stockholm University, Stockholm, Sweden.
Internet Interv. 2025 Jun 3;41:100841. doi: 10.1016/j.invent.2025.100841. eCollection 2025 Sep.
The use of artificial intelligence for psychological advice shows promise for enhancing accessibility and reducing costs, but it remains unclear whether AI-generated advice can match the quality and empathy of experts.
In a blinded, comparative cross-sectional design, licensed psychologists and psychotherapists assessed the quality, empathy, and authorship of psychological advice, which was either AI-generated or authored by experts.
AI-generated responses were rated significantly more favorable for emotional (OR = 1.79, 95 % CI [1.1, 2.93], = .02) and motivational empathy (OR = 1.84, 95 % CI [1.12, 3.04], = .02). Ratings for scientific quality ( = .10) and cognitive empathy ( = .08) were comparable to expert advice. Participants could not distinguish between AI- and expert-authored advice ( = .27), but expert authorship was associated with more favorable ratings across these measures (ORs for perceived AI vs. perceived expert ranging from 0.03 to 0.15, all < .001). For overall preference, AI-authored advice was favored when assessed blindly based on its actual source ( = 6.96, = .002). Nevertheless, advice as expert-authored was also strongly preferred ( = 6.26, = .001), with 93.55 % of participants preferring the advice they believed came from an expert, irrespective of its true origin.
AI demonstrates potential to match expert performance in asynchronous written psychological advice, but biases favoring perceived expert authorship may hinder its broader acceptance. Mitigating these biases and evaluating AI's trustworthiness and empathy are important next steps for safe and effective integration of AI in clinical practice.
利用人工智能提供心理建议在提高可及性和降低成本方面显示出前景,但人工智能生成的建议是否能与专家的质量和同理心相匹配仍不清楚。
在一项双盲、比较性横断面设计中,持牌心理学家和心理治疗师评估了人工智能生成或专家撰写的心理建议的质量、同理心和作者身份。
人工智能生成的回复在情感(优势比=1.79,95%置信区间[1.1,2.93],P=0.02)和动机同理心(优势比=1.84,95%置信区间[1.12,3.04],P=0.02)方面的评分明显更有利。科学质量(P=0.10)和认知同理心(P=0.08)的评分与专家建议相当。参与者无法区分人工智能撰写和专家撰写的建议(P=0.27),但在这些指标上,专家撰写与更有利的评分相关(感知为人工智能与感知为专家的优势比范围为0.03至0.15,均P<0.001)。对于总体偏好,当根据实际来源进行盲评时,人工智能撰写的建议更受青睐(χ²=6.96,P=0.002)。然而,被视为专家撰写的建议也非常受欢迎(χ²=6.26,P=0.001),93.55%的参与者更喜欢他们认为来自专家的建议,而不管其真实来源如何。
人工智能在异步书面心理建议方面显示出与专家表现相匹配的潜力,但有利于感知到的专家作者身份的偏见可能会阻碍其更广泛的接受。减轻这些偏见并评估人工智能的可信度和同理心是将人工智能安全有效地整合到临床实践中的重要下一步。