Division of Geriatrics, Department of Medicine, Vanderbilt University Medical Center, Nashville, TN, USA; Department of Biomedical Informatics, Vanderbilt University Medical Center, Nashville, TN, USA.
Department of Computer Science, Vanderbilt University, Nashville, TN, USA.
J Am Med Dir Assoc. 2024 Oct;25(10):105178. doi: 10.1016/j.jamda.2024.105178. Epub 2024 Aug 3.
There are many myths regarding Alzheimer's disease (AD) that have been circulated on the internet, each exhibiting varying degrees of accuracy, inaccuracy, and misinformation. Large language models, such as ChatGPT, may be a valuable tool to help assess these myths for veracity and inaccuracy; however, they can induce misinformation as well.
This study assesses ChatGPT's ability to identify and address AD myths with reliable information.
We conducted a cross-sectional study of attending geriatric medicine clinicians' evaluation of ChatGPT (GPT 4.0) responses to 16 selected AD myths. We prompted ChatGPT to express its opinion on each myth and implemented a survey using REDCap to determine the degree to which clinicians agreed with the accuracy of each of ChatGPT's explanations. We also collected their explanations of any disagreements with ChatGPT's responses. We used a 5-category Likert-type scale with a score ranging from -2 to 2 to quantify clinicians' agreement in each aspect of the evaluation.
The clinicians (n = 10) were generally satisfied with ChatGPT's explanations. Among the 16 myths, the clinicians were generally satisfied with these explanations, with [mean (SD) score of 1.1(±0.3)]. Most clinicians selected "Agree" or "Strongly Agree" for each statement. Some statements obtained a small number of "Disagree" responses. There were no "Strongly Disagree" responses.
Most surveyed health care professionals acknowledged the potential value of ChatGPT in mitigating AD misinformation; however, the need for more refined and detailed explanations of the disease's mechanisms and treatments was highlighted.
关于阿尔茨海默病(AD),互联网上流传着许多说法,准确性、不准确性和错误信息的程度各不相同。大型语言模型,如 ChatGPT,可能是评估这些说法真实性和准确性的有用工具;但是,它们也可能会引入错误信息。
本研究评估 ChatGPT 识别和提供可靠信息以纠正 AD 误解的能力。
我们对 10 名参加老年医学临床医生进行了一项横断面研究,评估他们对 ChatGPT(GPT 4.0)对 16 个选定的 AD 神话的回应。我们提示 ChatGPT 表达对每个神话的看法,并使用 REDCap 进行调查,以确定临床医生对 ChatGPT 每个解释的准确性的认可程度。我们还收集了他们对与 ChatGPT 回答不一致的解释。我们使用 5 级 Likert 量表,评分范围从-2 到 2,以量化临床医生在评估每个方面的一致性。
临床医生(n=10)对 ChatGPT 的解释普遍满意。在 16 个神话中,临床医生对这些解释普遍满意,平均(SD)得分为 1.1(±0.3)。大多数临床医生对每个陈述选择“同意”或“强烈同意”。一些陈述获得了少数“不同意”的回复。没有“强烈不同意”的回复。
大多数接受调查的医疗保健专业人员承认 ChatGPT 在减轻 AD 错误信息方面具有潜在价值;但是,需要更详细和详细地解释疾病的机制和治疗方法。