Wu Gloria, Lee David A, Zhao Weichen, Wong Adrial, Sidhu Sahej
Department of Ophthalmology, University of California San Francisco, San Francisco, CA, United States.
Department of Ophthalmology, McGovern Medical School, University of Texas Health Science Center, Houston, TX, United States.
Front Ophthalmol (Lausanne). 2023 Nov 16;3:1260415. doi: 10.3389/fopht.2023.1260415. eCollection 2023.
Our study investigates ChatGPT and its ability to communicate with glaucoma patients.
We inputted eight glaucoma-related questions/topics found on the American Academy of Ophthalmology (AAO)'s website into ChatGPT. We used the Flesch-Kincaid test, Gunning Fog Index, SMOG Index, and Dale-Chall readability formula to evaluate the comprehensibility of its responses for patients. ChatGPT's answers were compared with those found on the AAO's website.
ChatGPT's responses required reading comprehension of a higher grade level (average = grade 12.5 ± 1.6) than that of the text on the AAO's website (average = 9.4 grade ± 3.5), (0.0384). For the eight responses, the key ophthalmic terms appeared 34 out of 86 times in the ChatGPT responses vs. 86 out of 86 times in the text on the AAO's website. The term "eye doctor" appeared once in the ChatGPT text, but the formal term "ophthalmologist" did not appear. The term "ophthalmologist" appears 26 times on the AAO's website. The word counts of the answers produced by ChatGPT and those on the AAO's website were similar ( = 0.571), with phrases of a homogenous length.
ChatGPT trains on the texts, phrases, and algorithms inputted by software engineers. As ophthalmologists, through our websites and journals, we should consider encoding the phrase "see an ophthalmologist". Our medical assistants should sit with patients during their appointments to ensure that the text is accurate and that they fully comprehend its meaning. ChatGPT is effective for providing general information such as definitions or potential treatment options for glaucoma. However, ChatGPT has a tendency toward repetitive answers and, due to their elevated readability scores, these could be too difficult for a patient to read.
我们的研究调查了ChatGPT及其与青光眼患者沟通的能力。
我们将在美国眼科学会(AAO)网站上找到的八个与青光眼相关的问题/主题输入ChatGPT。我们使用弗莱施-金凯德测试、冈宁雾度指数、烟雾指数和戴尔-查尔可读性公式来评估其回答对患者的可理解性。将ChatGPT的答案与AAO网站上的答案进行比较。
ChatGPT的回答所需的阅读理解年级水平(平均=12.5±1.6年级)高于AAO网站上的文本(平均=9.4±3.5年级),(P=0.0384)。对于这八个回答,关键眼科术语在ChatGPT的回答中出现了34次(共86次),而在AAO网站的文本中出现了86次(共86次)。“眼科医生”一词在ChatGPT的文本中出现了一次,但正式术语“眼科专家”未出现。“眼科专家”一词在AAO网站上出现了26次。ChatGPT生成的答案和AAO网站上的答案的单词数相似(P=0.571),短语长度均匀。
ChatGPT基于软件工程师输入的文本、短语和算法进行训练。作为眼科医生,我们应该通过我们的网站和期刊考虑编入“看眼科专家”这一短语。我们的医疗助理在患者预约就诊时应与患者一起,以确保文本准确且患者完全理解其含义。ChatGPT在提供一般信息(如青光眼的定义或潜在治疗方案)方面是有效的。然而,ChatGPT有重复回答的倾向,并且由于其较高的可读性分数,这些回答对患者来说可能太难读懂。