Suppr超能文献

基于ChatGPT-4的斜视和弱视常见问题回答的可接受性与可读性。

Acceptability and readability of ChatGPT-4 based responses for frequently asked questions about strabismus and amblyopia.

作者信息

Guven S, Ayyildiz B

机构信息

Kayseri City Hospital, Department of Ophthalmology, Kayseri, Turkey.

Kayseri City Hospital, Department of Ophthalmology, Kayseri, Turkey.

出版信息

J Fr Ophtalmol. 2025 Mar;48(3):104400. doi: 10.1016/j.jfo.2024.104400. Epub 2024 Dec 20.

Abstract

PURPOSE

To evaluate the compatibility and readability of ChatGPT-4 in providing responses to common inquiries about strabismus and amblyopia.

MATERIALS AND METHODS

A series of commonly asked questions were compiled, covering topics such as the definition, prevalence, diagnostic approaches, surgical and non-surgical treatment alternatives, postoperative guidelines, surgery-related risks, and visual prognosis associated with strabismus and amblyopia. Each question was asked three times on the online ChatGPT-4 platform both in English and French, with data collection conducted on February 18, 2024. The responses generated by ChatGPT-4 were evaluated by two independent pediatric ophthalmologists, who classified them as "acceptable," "unacceptable," or "incomplete." Additionally, an online readability assessment tool called "readable" was utilized for readability analysis.

RESULTS

The majority of responses, totaling 97% of the questions regarding strabismus and amblyopia, consistently met the criteria for acceptability. Only 3% of responses were classified as incomplete, with no instances of unacceptable responses observed. The average Flesch-Kincaid Grade Level and Flesch Reading Ease Score were calculated as 14.53±1.8 and 23.63±8.2, respectively. Furthermore, the means for all readability indices, including the Coleman-Liau index, the Gunning Fog index, and the SMOG index, were found to be 15.75±1.4, 16.96±2.4, and 16.05±1.6, respectively.

CONCLUSIONS

ChatGPT-4 consistently produced acceptable responses to the majority of the questions asked (97%). Nevertheless, the readability of these responses proved challenging for the average layperson, requiring a college-level education for comprehension. Further improvements, particularly in terms of readability, are necessary to enhance the advisory capacity of this AI software in providing eye and health-related guidance for patients, physicians, and the general public.

摘要

目的

评估ChatGPT-4对有关斜视和弱视常见问题提供回答的兼容性和可读性。

材料与方法

汇编了一系列常见问题,涵盖斜视和弱视的定义、患病率、诊断方法、手术和非手术治疗选择、术后指南、手术相关风险以及视觉预后等主题。2024年2月18日,在在线ChatGPT-4平台上用英语和法语对每个问题各提问三次,并进行数据收集。由两名独立的儿科眼科医生对ChatGPT-4生成的回答进行评估,他们将回答分为“可接受”、“不可接受”或“不完整”。此外,还使用了一个名为“readable”的在线可读性评估工具进行可读性分析。

结果

关于斜视和弱视的问题,大多数回答(占97%)始终符合可接受标准。只有3%的回答被归类为不完整,未观察到不可接受的回答。平均弗莱施-金凯德年级水平和弗莱施阅读易度得分分别计算为14.53±1.8和23.63±8.2。此外,所有可读性指标的平均值,包括科尔曼-廖指数、冈宁雾度指数和烟雾指数,分别为15.75±1.4、16.96±2.4和16.05±1.6。

结论

ChatGPT-4对大多数提出的问题(97%)始终给出可接受的回答。然而,这些回答的可读性对普通外行人来说具有挑战性,需要大学水平的教育才能理解。有必要进一步改进,特别是在可读性方面,以提高该人工智能软件为患者、医生和公众提供眼睛和健康相关指导的咨询能力。

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验