Huang Sean S, Song Qingyuan, Beiting Kimberly J, Duggan Maria C, Hines Kristin, Murff Harvey, Leung Vania, Powers James, Harvey T S, Malin Bradley, Yin Zhijun
Department of Medicine, Division of Geriatrics, Vanderbilt University Medical Center, Nashville, TN, USA.
Department of Biomedical Informatics, Vanderbilt University Medical Center, Nashville, TN, USA.
medRxiv. 2023 Sep 7:2023.09.04.23294917. doi: 10.1101/2023.09.04.23294917.
There are many myths regarding Alzheimer's disease (AD) that have been circulated on the Internet, each exhibiting varying degrees of accuracy, inaccuracy, and misinformation. Large language models such as ChatGPT, may be a useful tool to help assess these myths for veracity and inaccuracy. However, they can induce misinformation as well. The objective of this study is to assess ChatGPT's ability to identify and address AD myths with reliable information.
We conducted a cross-sectional study of clinicians' evaluation of ChatGPT (GPT 4.0)'s responses to 20 selected AD myths. We prompted ChatGPT to express its opinion on each myth and then requested it to rephrase its explanation using a simplified language that could be more readily understood by individuals with a middle school education. We implemented a survey using Redcap to determine the degree to which clinicians agreed with the accuracy of each ChatGPT's explanation and the degree to which the simplified rewriting was readable and retained the message of the original. We also collected their explanation on any disagreement with ChatGPT's responses. We used five Likert-type scale with a score ranging from -2 to 2 to quantify clinicians' agreement in each aspect of the evaluation.
The clinicians (n=11) were generally satisfied with ChatGPT's explanations, with a mean (SD) score of 1.0(±0.3) across the 20 myths. While ChatGPT correctly identified that all the 20 myths were inaccurate, some clinicians disagreed with its explanations on 7 of the myths.Overall, 9 of the 11 professionals either agreed or strongly agreed that ChatGPT has the potential to provide meaningful explanations of certain myths.
The majority of surveyed healthcare professionals acknowledged the potential value of ChatGPT in mitigating AD misinformation. However, the need for more refined and detailed explanations of the disease's mechanisms and treatments was highlighted.
互联网上流传着许多关于阿尔茨海默病(AD)的谣言,其准确性、不准确性和错误信息程度各不相同。像ChatGPT这样的大型语言模型可能是帮助评估这些谣言真伪的有用工具。然而,它们也可能引发错误信息。本研究的目的是评估ChatGPT识别并以可靠信息应对AD谣言的能力。
我们进行了一项横断面研究,让临床医生评估ChatGPT(GPT 4.0)对20个选定的AD谣言的回应。我们促使ChatGPT对每个谣言发表意见,然后要求它用更简单的语言重新表述其解释,以便初中文化程度的人更容易理解。我们使用Redcap进行了一项调查,以确定临床医生对ChatGPT每个解释准确性的认同程度,以及简化重写后的内容可读性和保留原文信息的程度。我们还收集了他们对与ChatGPT回应存在分歧之处的解释。我们使用了一个从-2到2的五点李克特量表来量化临床医生在评估各方面的认同程度。
临床医生(n = 11)总体上对ChatGPT的解释感到满意,20个谣言的平均(标准差)评分为1.0(±0.3)。虽然ChatGPT正确识别出所有20个谣言都是不准确的,但一些临床医生对其中7个谣言的解释存在分歧。总体而言,11名专业人员中有9人同意或强烈同意ChatGPT有潜力对某些谣言提供有意义的解释。
大多数接受调查的医疗保健专业人员认可ChatGPT在减少AD错误信息方面的潜在价值。然而,强调了对该疾病机制和治疗方法进行更精确、详细解释的必要性。