Tianjin University of Traditional Chinese Medicine, Tianjin, China.
Dongguan Rehabilitation Experimental School, Dongguan, China.
J Med Internet Res. 2024 Apr 30;26:e54706. doi: 10.2196/54706.
BACKGROUND: There is a dearth of feasibility assessments regarding using large language models (LLMs) for responding to inquiries from autistic patients within a Chinese-language context. Despite Chinese being one of the most widely spoken languages globally, the predominant research focus on applying these models in the medical field has been on English-speaking populations. OBJECTIVE: This study aims to assess the effectiveness of LLM chatbots, specifically ChatGPT-4 (OpenAI) and ERNIE Bot (version 2.2.3; Baidu, Inc), one of the most advanced LLMs in China, in addressing inquiries from autistic individuals in a Chinese setting. METHODS: For this study, we gathered data from DXY-a widely acknowledged, web-based, medical consultation platform in China with a user base of over 100 million individuals. A total of 100 patient consultation samples were rigorously selected from January 2018 to August 2023, amounting to 239 questions extracted from publicly available autism-related documents on the platform. To maintain objectivity, both the original questions and responses were anonymized and randomized. An evaluation team of 3 chief physicians assessed the responses across 4 dimensions: relevance, accuracy, usefulness, and empathy. The team completed 717 evaluations. The team initially identified the best response and then used a Likert scale with 5 response categories to gauge the responses, each representing a distinct level of quality. Finally, we compared the responses collected from different sources. RESULTS: Among the 717 evaluations conducted, 46.86% (95% CI 43.21%-50.51%) of assessors displayed varying preferences for responses from physicians, with 34.87% (95% CI 31.38%-38.36%) of assessors favoring ChatGPT and 18.27% (95% CI 15.44%-21.10%) of assessors favoring ERNIE Bot. The average relevance scores for physicians, ChatGPT, and ERNIE Bot were 3.75 (95% CI 3.69-3.82), 3.69 (95% CI 3.63-3.74), and 3.41 (95% CI 3.35-3.46), respectively. Physicians (3.66, 95% CI 3.60-3.73) and ChatGPT (3.73, 95% CI 3.69-3.77) demonstrated higher accuracy ratings compared to ERNIE Bot (3.52, 95% CI 3.47-3.57). In terms of usefulness scores, physicians (3.54, 95% CI 3.47-3.62) received higher ratings than ChatGPT (3.40, 95% CI 3.34-3.47) and ERNIE Bot (3.05, 95% CI 2.99-3.12). Finally, concerning the empathy dimension, ChatGPT (3.64, 95% CI 3.57-3.71) outperformed physicians (3.13, 95% CI 3.04-3.21) and ERNIE Bot (3.11, 95% CI 3.04-3.18). CONCLUSIONS: In this cross-sectional study, physicians' responses exhibited superiority in the present Chinese-language context. Nonetheless, LLMs can provide valuable medical guidance to autistic patients and may even surpass physicians in demonstrating empathy. However, it is crucial to acknowledge that further optimization and research are imperative prerequisites before the effective integration of LLMs in clinical settings across diverse linguistic environments can be realized. TRIAL REGISTRATION: Chinese Clinical Trial Registry ChiCTR2300074655; https://www.chictr.org.cn/bin/project/edit?pid=199432.
背景:在中文语境下,使用大型语言模型(LLM)来回应自闭症患者的咨询,可行性评估还很少。尽管中文是全球使用最广泛的语言之一,但这些模型在医学领域的应用主要集中在英语人群上。
目的:本研究旨在评估 LLM 聊天机器人,特别是 ChatGPT-4(OpenAI)和 ERNIE Bot(百度公司,版本 2.2.3)在中国,在处理中文环境下自闭症个体咨询方面的有效性。
方法:在这项研究中,我们从中国知名的在线医疗咨询平台 DXY 上收集了数据,该平台拥有超过 1 亿用户。从 2018 年 1 月至 2023 年 8 月,我们严格选择了 100 名患者咨询样本,从平台上公开的自闭症相关文档中提取了 239 个问题。为了保持客观性,我们对原始问题和回答进行了匿名和随机化处理。一个由 3 名主治医生组成的评估小组对 4 个维度的相关性、准确性、有用性和同理心进行了评估:相关性、准确性、有用性和同理心。评估小组共完成了 717 次评估。团队最初确定了最佳回答,然后使用了 5 个反应类别进行评估,每个类别代表不同的质量水平。最后,我们比较了不同来源的反应。
结果:在进行的 717 次评估中,46.86%(95%置信区间 43.21%-50.51%)的评估者对医生的回答有不同的偏好,其中 34.87%(95%置信区间 31.38%-38.36%)的评估者更喜欢 ChatGPT,18.27%(95%置信区间 15.44%-21.10%)的评估者更喜欢 ERNIE Bot。医生、ChatGPT 和 ERNIE Bot 的平均相关性评分分别为 3.75(95%置信区间 3.69-3.82)、3.69(95%置信区间 3.63-3.74)和 3.41(95%置信区间 3.35-3.46)。医生(3.66,95%置信区间 3.60-3.73)和 ChatGPT(3.73,95%置信区间 3.69-3.77)的准确性评分高于 ERNIE Bot(3.52,95%置信区间 3.47-3.57)。在有用性评分方面,医生(3.54,95%置信区间 3.47-3.62)的评分高于 ChatGPT(3.40,95%置信区间 3.34-3.47)和 ERNIE Bot(3.05,95%置信区间 2.99-3.12)。最后,在同理心维度方面,ChatGPT(3.64,95%置信区间 3.57-3.71)优于医生(3.13,95%置信区间 3.04-3.21)和 ERNIE Bot(3.11,95%置信区间 3.04-3.18)。
结论:在这项横断面研究中,医生的回答在目前的中文语境下具有优势。然而,LLM 可以为自闭症患者提供有价值的医疗指导,甚至在表现出同理心方面可能超过医生。但是,必须认识到,在不同语言环境下,有效的整合到临床环境中之前,还需要进行进一步的优化和研究。
试验注册:中国临床试验注册中心 ChiCTR2300074655;https://www.chictr.org.cn/bin/project/edit?pid=199432.
J Med Internet Res. 2025-6-4
J Med Internet Res. 2025-7-22
J Med Internet Res. 2024-12-11
Zhonghua Kou Qiang Yi Xue Za Zhi. 2025-7-30
J Cancer Educ. 2025-5-26
J Med Internet Res. 2025-5-5
J Am Med Inform Assoc. 2025-5-1
J Clin Med. 2025-1-28
JMIR Ment Health. 2024-10-18
Ann Intern Med. 2024-2
Brief Bioinform. 2023-11-22