Keyßer Gernot, Pfeil Alexander, Reuß-Borst Monika, Frohne Inna, Schultz Olaf, Sander Oliver
Klinik und Poliklinik für Innere Medizin II, Universitätsklinikum Halle, Ernst-Grube-Str. 40, 06120, Halle (Saale), Deutschland.
Klinik für Innere Medizin III, Universitätsklinikum Jena, Friedrich-Schiller-Universität Jena, Jena, Deutschland.
Z Rheumatol. 2025 Apr;84(3):179-187. doi: 10.1007/s00393-024-01535-6. Epub 2024 Jul 10.
INTRODUCTION: The chatbot ChatGPT represents a milestone in the interaction between humans and large databases that are accessible via the internet. It facilitates the answering of complex questions by enabling a communication in everyday language. Therefore, it is a potential source of information for those who are affected by rheumatic diseases. The aim of our investigation was to find out whether ChatGPT (version 3.5) is capable of giving qualified answers regarding the application of specific methods of complementary and alternative medicine (CAM) in three rheumatic diseases: rheumatoid arthritis (RA), systemic lupus erythematosus (SLE) and granulomatosis with polyangiitis (GPA). In addition, it was investigated how the answers of the chatbot were influenced by the wording of the question. METHODS: The questioning of ChatGPT was performed in three parts. Part A consisted of an open question regarding the best way of treatment of the respective disease. In part B, the questions were directed towards possible indications for the application of CAM in general in one of the three disorders. In part C, the chatbot was asked for specific recommendations regarding one of three CAM methods: homeopathy, ayurvedic medicine and herbal medicine. Questions in parts B and C were expressed in two modifications: firstly, it was asked whether the specific CAM was applicable at all in certain rheumatic diseases. The second question asked which procedure of the respective CAM method worked best in the specific disease. The validity of the answers was checked by using the ChatGPT reliability score, a Likert scale ranging from 1 (lowest validity) to 7 (highest validity). RESULTS: The answers to the open questions of part A had the highest validity. In parts B and C, ChatGPT suggested a variety of CAM applications that lacked scientific evidence. The validity of the answers depended on the wording of the questions. If the question suggested the inclination to apply a certain CAM, the answers often lacked the information of missing evidence and were graded with lower score values. CONCLUSION: The answers of ChatGPT (version 3.5) regarding the applicability of CAM in selected rheumatic diseases are not convincingly based on scientific evidence. In addition, the wording of the questions affects the validity of the information. Currently, an uncritical application of ChatGPT as an instrument for patient information cannot be recommended.
引言:聊天机器人ChatGPT代表了人类与可通过互联网访问的大型数据库之间交互的一个里程碑。它通过实现日常语言的交流来促进复杂问题的解答。因此,它对于受风湿性疾病影响的人来说是一个潜在的信息来源。我们调查的目的是了解ChatGPT(3.5版本)是否能够就补充和替代医学(CAM)的特定方法在三种风湿性疾病:类风湿性关节炎(RA)、系统性红斑狼疮(SLE)和肉芽肿性多血管炎(GPA)中的应用给出合格的答案。此外,还研究了聊天机器人的答案如何受到问题措辞的影响。 方法:对ChatGPT的提问分三个部分进行。A部分包括一个关于各自疾病最佳治疗方法的开放性问题。在B部分,问题针对的是在这三种疾病之一中一般应用CAM的可能适应症。在C部分,要求聊天机器人就三种CAM方法之一:顺势疗法、阿育吠陀医学和草药医学给出具体建议。B部分和C部分的问题有两种表述方式:首先,询问特定的CAM在某些风湿性疾病中是否根本适用。第二个问题是询问各自CAM方法的哪种程序在特定疾病中效果最佳。通过使用ChatGPT可靠性评分来检查答案的有效性,这是一个从1(最低有效性)到7(最高有效性)的李克特量表。 结果:A部分开放性问题的答案有效性最高。在B部分和C部分,ChatGPT提出了各种缺乏科学证据的CAM应用。答案的有效性取决于问题的措辞。如果问题暗示了应用某种CAM的倾向,答案往往缺乏缺乏证据的信息,并且得分较低。 结论:ChatGPT(3.5版本)关于CAM在选定风湿性疾病中的适用性的答案并非令人信服地基于科学证据。此外,问题的措辞会影响信息的有效性。目前,不建议不加批判地将ChatGPT用作患者信息工具。
Int J Rheum Dis. 2023-7
J Shoulder Elbow Surg. 2025-5
Tech Coloproctol. 2025-1-26
Hand Surg Rehabil. 2025-4
J Thorac Cardiovasc Surg. 2025-4
Patient Prefer Adherence. 2025-7-31
Genes (Basel). 2024-12-27
Genes (Basel). 2024-3-28
Lancet Rheumatol. 2024-4
J Orthop Translat. 2024-2-6
Proc Natl Acad Sci U S A. 2024-2-27
Nurse Educ Today. 2024-4