Karve Zev, Calpey Jacob, Machado Christopher, Knecht Michelle, Mejia Maria Carmenza
Schmidt College of Medicine, Florida Atlantic University, Boca Raton, FL, United States.
J Med Internet Res. 2025 Sep 16;27:e78417. doi: 10.2196/78417.
Artificial intelligence (AI) is increasingly used in digital health, particularly through large language models (LLMs), to support patient engagement and behavior change. One novel application is the delivery of motivational interviewing (MI), an evidence-based, patient-centered counseling technique designed to enhance motivation and resolve ambivalence around health behaviors. AI tools, including chatbots, mobile apps, and web-based agents, are being developed to simulate MI techniques at scale. While these innovations are promising, important questions remain about how faithfully AI systems can replicate MI principles or achieve meaningful behavioral impact.
This scoping review aimed to summarize existing empirical studies evaluating AI-driven systems that apply MI techniques to support health behavior change. Specifically, we examined the feasibility of these systems; their fidelity to MI principles; and their reported behavioral, psychological, or engagement outcomes.
We systematically searched PubMed, Embase, Scopus, Web of Science, and Cochrane Library for empirical studies published between January 1, 2018, and February 25, 2025. Eligible studies involved AI-driven systems using natural language generation, understanding, or computational logic to deliver MI techniques to users targeting a specific health behavior. We excluded studies using AI solely for training clinicians in MI. Three independent reviewers screened and extracted data on study design, AI modality and type, MI components, health behavior focus, MI fidelity assessment, and outcome domains.
Of the 1001 records identified, 15 (1.5%) met the inclusion criteria. Of these 15 studies, 6 (40%) were exploratory feasibility or pilot studies, and 3 (20%) were randomized controlled trials. AI modalities included rule-based chatbots (9/15, 60%), LLM-based systems (4/15, 27%), and virtual or mobile agents (2/15, 13%). Targeted behaviors included smoking cessation (6/15, 40%), substance use (3/15, 20%), COVID-19 vaccine hesitancy, type 2 diabetes self-management, stress, mental health service use, and opioid use during pregnancy. Of the 15 studies, 13 (87%) reported positive findings on feasibility or user acceptability, while 6 (40%) assessed MI fidelity using expert review or structured coding, with moderate to high alignment reported. Several studies found that users perceived the AI systems as judgment free, supportive, and easier to engage with than human counselors, particularly in stigmatized contexts. However, limitations in empathy, safety transparency, and emotional nuance were commonly noted. Only 3 (20%) of the 15 studies reported substantially significant behavioral changes.
AI systems delivering MI show promise for enhancing patient engagement and scaling behavior change interventions. Early evidence supports their usability and partial fidelity to MI principles, especially in sensitive domains. However, most systems remain in early development, and few have been rigorously tested. Future research should prioritize randomized evaluations; standardized fidelity measures; and safeguards for LLM safety, empathy, and accuracy in health-related dialogue.
OSF Registries 10.17605/OSF.IO/G9N7E; https://osf.io/g9n7e.
人工智能(AI)在数字健康领域的应用日益广泛,特别是通过大语言模型(LLM)来支持患者参与和行为改变。一种新颖的应用是提供动机性访谈(MI),这是一种基于证据、以患者为中心的咨询技术,旨在增强动机并解决围绕健康行为的矛盾心理。包括聊天机器人、移动应用程序和基于网络的智能体在内的人工智能工具正在被开发,以大规模模拟动机性访谈技术。虽然这些创新很有前景,但关于人工智能系统能够多忠实地复制动机性访谈原则或实现有意义的行为影响,仍存在重要问题。
本范围综述旨在总结现有实证研究,这些研究评估了应用动机性访谈技术来支持健康行为改变的人工智能驱动系统。具体而言,我们考察了这些系统的可行性;它们对动机性访谈原则的忠诚度;以及它们报告的行为、心理或参与结果。
我们系统地检索了PubMed、Embase、Scopus、Web of Science和Cochrane图书馆,以查找2018年1月1日至2025年2月25日期间发表的实证研究。符合条件的研究涉及使用自然语言生成、理解或计算逻辑向针对特定健康行为的用户提供动机性访谈技术的人工智能驱动系统。我们排除了仅将人工智能用于培训临床医生进行动机性访谈的研究。三位独立评审员筛选并提取了关于研究设计、人工智能模式和类型、动机性访谈组成部分、健康行为重点、动机性访谈忠诚度评估和结果领域的数据。
在识别出的1001条记录中,15条(1.5%)符合纳入标准。在这15项研究中,6项(40%)是探索性可行性研究或试点研究,3项(20%)是随机对照试验。人工智能模式包括基于规则的聊天机器人(9/15,60%)、基于大语言模型的系统(4/15,27%)以及虚拟或移动智能体(2/15,13%)。目标行为包括戒烟(6/15,40%)、物质使用(3/15,20%)、对新冠疫苗的犹豫态度、2型糖尿病自我管理、压力、心理健康服务使用以及孕期阿片类药物使用。在这15项研究中,13项(87%)报告了关于可行性或用户可接受性的积极结果,而6项(40%)使用专家评审或结构化编码评估了动机性访谈的忠诚度,报告显示一致性为中度到高度。几项研究发现,用户认为人工智能系统没有评判性、具有支持性,并且比人类咨询师更容易接触,特别是在有污名化的情境中。然而,共情、安全透明度和情感细微差别方面的局限性普遍被提及。15项研究中只有3项(20%)报告了显著的行为变化。
提供动机性访谈的人工智能系统在增强患者参与和扩大行为改变干预方面显示出前景。早期证据支持它们的可用性以及对动机性访谈原则的部分忠诚度,特别是在敏感领域。然而,大多数系统仍处于早期开发阶段,很少经过严格测试。未来的研究应优先进行随机评估;采用标准化的忠诚度测量方法;并对大语言模型在健康相关对话中的安全性、共情能力和准确性采取保障措施。
OSF注册库10.17605/OSF.IO/G9N7E;https://osf.io/g9n7e 。