Guo Weiqi, Chen Yang
School of Foreign Languages, Renmin University of China, Beijing, China.
School of Journalism and Communication, Renmin University of China, Beijing, China.
J Med Internet Res. 2025 Mar 5;27:e66760. doi: 10.2196/66760.
The increasing use of artificial intelligence (AI) in medical diagnosis and consultation promises benefits such as greater accuracy and efficiency. However, there is little evidence to systematically test whether the ideal technological promises translate into an improved evaluation of the medical consultation from the patient's perspective. This perspective is significant because AI as a technological solution does not necessarily improve patient confidence in diagnosis and adherence to treatment at the functional level, create meaningful interactions between the medical agent and the patient at the relational level, evoke positive emotions, or reduce the patient's pessimism at the emotional level.
This study aims to investigate, from a patient-centered perspective, whether AI or human-involved AI can replace the role of human physicians in diagnosis at the functional, relational, and emotional levels as well as how some health-related differences between human-AI and human-human interactions affect patients' evaluations of the medical consultation.
A 3 (consultation source: AI vs human-involved AI vs human) × 2 (health-related stigma: low vs high) × 2 (diagnosis explanation: without vs with explanation) factorial experiment was conducted with 249 participants. The main effects and interaction effects of the variables were examined on individuals' functional, relational, and emotional evaluations of the medical consultation.
Functionally, people trusted the diagnosis of the human physician (mean 4.78-4.85, SD 0.06-0.07) more than medical AI (mean 4.34-4.55, SD 0.06-0.07) or human-involved AI (mean 4.39-4.56, SD 0.06-0.07; P<.001), but at the relational and emotional levels, there was no significant difference between human-AI and human-human interactions (P>.05). Health-related stigma had no significant effect on how people evaluated the medical consultation or contributed to preferring AI-powered systems over humans (P>.05); however, providing explanations of the diagnosis significantly improved the functional (P<.001), relational (P<.05), and emotional (P<.05) evaluations of the consultation for all 3 medical agents.
The findings imply that at the current stage of AI development, people trust human expertise more than accurate AI, especially for decisions traditionally made by humans, such as medical diagnosis, supporting the algorithm aversion theory. Surprisingly, even for highly stigmatized diseases such as AIDS, where we assume anonymity and privacy are preferred in medical consultations, the dehumanization of AI does not contribute significantly to the preference for AI-powered medical agents over humans, suggesting that instrumental needs of diagnosis override patient privacy concerns. Furthermore, explaining the diagnosis effectively improves treatment adherence, strengthens the physician-patient relationship, and fosters positive emotions during the consultation. This provides insights for the design of AI medical agents, which have long been criticized for lacking transparency while making highly consequential decisions. This study concludes by outlining theoretical contributions to research on health communication and human-AI interaction and discusses the implications for the design and application of medical AI.
人工智能(AI)在医学诊断和咨询中的应用日益广泛,有望带来更高的准确性和效率等益处。然而,几乎没有证据能系统地检验这些理想的技术承诺是否能从患者的角度转化为对医学咨询的改进评估。这一观点很重要,因为作为一种技术解决方案,人工智能不一定能在功能层面提高患者对诊断的信心和对治疗的依从性,在关系层面在医疗服务提供者与患者之间建立有意义的互动,唤起积极情绪,或在情感层面减轻患者的悲观情绪。
本研究旨在从以患者为中心的角度调查人工智能或人机协作的人工智能是否能在功能、关系和情感层面取代人类医生在诊断中的作用,以及人机交互与人人交互之间一些与健康相关的差异如何影响患者对医学咨询的评估。
对249名参与者进行了一项3(咨询来源:人工智能 vs 人机协作的人工智能 vs 人类)×2(与健康相关的污名:低 vs 高)×2(诊断解释:无 vs 有解释)的析因实验。研究了这些变量对个体对医学咨询的功能、关系和情感评估的主效应和交互效应。
在功能上,人们对人类医生诊断的信任度(均值4.78 - 4.85,标准差0.06 - 0.07)高于医学人工智能(均值4.34 - 4.55,标准差0.06 - 0.07)或人机协作的人工智能(均值4.39 - 4.56,标准差0.06 - 0.07;P <.001),但在关系和情感层面,人机交互与人人交互之间没有显著差异(P >.05)。与健康相关的污名对人们如何评估医学咨询或导致倾向于人工智能驱动的系统而非人类没有显著影响(P >.05);然而,提供诊断解释显著改善了所有3种医疗服务提供者的咨询在功能(P <.001)、关系(P <.05)和情感(P <.05)方面的评估。
研究结果表明,在人工智能发展的现阶段,人们更信任人类专业知识而非精确的人工智能,尤其是对于传统上由人类做出的决策,如医学诊断,这支持了算法厌恶理论。令人惊讶的是,即使对于像艾滋病这样高度污名化的疾病,我们假设在医学咨询中匿名和隐私更受青睐,但人工智能的非人性化并没有显著导致更倾向于人工智能驱动的医疗服务提供者而非人类,这表明诊断的工具性需求优先于患者的隐私担忧。此外,解释诊断有效地提高了治疗依从性,加强了医患关系,并在咨询过程中促进了积极情绪。这为长期以来因在做出重大决策时缺乏透明度而受到批评的人工智能医疗服务提供者的设计提供了见解。本研究最后概述了对健康传播和人机交互研究的理论贡献,并讨论了对医学人工智能设计和应用的启示。