• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

关注文本:评估人工智能和眼科医生对患者手术疑问的回复的可读性。

Eyes on the Text: Assessing Readability of Artificial Intelligence and Ophthalmologist Responses to Patient Surgery Queries.

作者信息

Kurapati Sai S, Barnett Derek J, Yaghy Antonio, Sabet Cameron J, Younessi David N, Nguyen Dang, Lin John C, Scott Ingrid U

机构信息

Department of Ophthalmology, Penn State College of Medicine, Hershey, Pennsylvania, USA,

Department of Ophthalmology, Penn State College of Medicine, Hershey, Pennsylvania, USA.

出版信息

Ophthalmologica. 2025;248(3):149-159. doi: 10.1159/000544917. Epub 2025 Mar 10.

DOI:10.1159/000544917
PMID:40064147
Abstract

INTRODUCTION

Generative artificial intelligence (AI) technologies like GPT-4 can instantaneously provide health information to patients; however, the readability of these outputs compared to ophthalmologist-written responses is unknown. This study aimed to evaluate the readability of GPT-4-generated and ophthalmologist-written responses to patient queries about ophthalmic surgery.

METHODS

This retrospective cross-sectional study used 200 randomly selected patient questions about ophthalmic surgery extracted from the American Academy of Ophthalmology's EyeSmart platform. The questions were inputted into GPT-4, and the generated responses were recorded. Ophthalmologist-written replies to the same questions were compiled for comparison. Readability of GPT-4 and ophthalmologist responses was assessed using six validated metrics: Flesch Kincaid Reading Ease (FK-RE), Flesch Kincaid Grade Level (FK-GL), Gunning Fog Score (GFS), SMOG Index (SI), Coleman Liau Index (CLI), and Automated Readability Index (ARI). Descriptive statistics, one-way ANOVA, Shapiro-Wilk, and Levene's tests (α = 0.05) were used to compare readability between the two groups.

RESULTS

GPT-4 used a higher percentage of complex words (24.42%) compared to ophthalmologists (17.76%), although mean (standard deviation) word count per sentence was similar (18.43 [2.95] and 18.01 [6.09]). Across all metrics (FK-RE; FK-GL; GFS; SI; CLI; and ARI), GPT-4 responses were at a higher grade level (34.39 [8.51]; 13.19 [2.63]; 16.37 [2.04]; 12.18 [1.43]; 15.72 [1.40]; 12.99 [1.86]) than ophthalmologists' responses (50.61 [15.53]; 10.71 [2.99]; 14.13 [3.55]; 10.07 [2.46]; 12.64 [2.93]; 10.40 [3.61]), with both sources necessitating a 12th-grade education for comprehension. ANOVA tests showed significance (p < 0.05) for all comparisons except word count (p = 0.438).

CONCLUSION

The National Institutes of Health advises health information to be written at a 6th- to 7th-grade level. Both GPT-4- and ophthalmologist-written answers exceeded this recommendation, with GPT-4 showing a greater gap. Information accessibility is vital when designing patient resources, particularly with the rise of AI as an educational tool.

摘要

引言

像GPT-4这样的生成式人工智能(AI)技术可以即时向患者提供健康信息;然而,与眼科医生撰写的回复相比,这些输出内容的可读性尚不清楚。本研究旨在评估GPT-4生成的以及眼科医生撰写的针对患者有关眼科手术问题的回复的可读性。

方法

这项回顾性横断面研究使用了从美国眼科学会的EyeSmart平台随机抽取的200个关于眼科手术的患者问题。将这些问题输入GPT-4,并记录生成的回复。收集眼科医生对相同问题的书面回复以作比较。使用六个经过验证的指标评估GPT-4和眼科医生回复的可读性:弗莱什-金凯德阅读简易度(FK-RE)、弗莱什-金凯德年级水平(FK-GL)、冈宁雾度评分(GFS)、烟雾指数(SI)、科尔曼-廖指数(CLI)和自动可读性指数(ARI)。使用描述性统计、单因素方差分析、夏皮罗-威尔克检验和莱文检验(α = 0.05)比较两组之间的可读性。

结果

与眼科医生(17.76%)相比,GPT-4使用的复杂词汇百分比更高(24.42%),尽管平均每句单词数(标准差)相似(18.43 [2.95]和18.01 [6.09])。在所有指标(FK-RE;FK-GL;GFS;SI;CLI;和ARI)上,GPT-4的回复年级水平更高(34.39 [8.51];13.19 [2.63];16.37 [2.04];12.18 [1.43];15.72 [1.40];12.99 [1.86]),高于眼科医生的回复(50.61 [15.53];10.71 [2.99];14.13 [3.55];10.07 [2.46];12.64 [2.93];10.40 [3.61]),理解这两种来源的回复都需要十二年级的教育水平。方差分析测试显示,除单词数外(p = 0.438),所有比较均具有显著性(p < 0.05)。

结论

美国国立卫生研究院建议健康信息应写成六年级至七年级的水平。GPT-4和眼科医生撰写的答案都超出了这一建议,GPT-4的差距更大。在设计患者资源时,信息的可获取性至关重要,尤其是随着人工智能作为一种教育工具的兴起。

相似文献

1
Eyes on the Text: Assessing Readability of Artificial Intelligence and Ophthalmologist Responses to Patient Surgery Queries.关注文本:评估人工智能和眼科医生对患者手术疑问的回复的可读性。
Ophthalmologica. 2025;248(3):149-159. doi: 10.1159/000544917. Epub 2025 Mar 10.
2
Artificial Intelligence in Peripheral Artery Disease Education: A Battle Between ChatGPT and Google Gemini.外周动脉疾病教育中的人工智能:ChatGPT与谷歌Gemini的较量
Cureus. 2025 Jun 1;17(6):e85174. doi: 10.7759/cureus.85174. eCollection 2025 Jun.
3
Enhancing the Readability of Online Patient Education Materials Using Large Language Models: Cross-Sectional Study.使用大语言模型提高在线患者教育材料的可读性:横断面研究。
J Med Internet Res. 2025 Jun 4;27:e69955. doi: 10.2196/69955.
4
Artificial Intelligence Shows Limited Success in Improving Readability Levels of Spanish-language Orthopaedic Patient Education Materials.人工智能在提高西班牙语骨科患者教育材料的可读性方面成效有限。
Clin Orthop Relat Res. 2025 Feb 11. doi: 10.1097/CORR.0000000000003413.
5
Readability analysis as a tool for evaluating English proficiency in first-year medical students.可读性分析作为评估一年级医学生英语水平的一种工具。
BMC Med Educ. 2025 Jul 1;25(1):945. doi: 10.1186/s12909-025-07348-8.
6
American Academy of Orthopaedic Surgeons OrthoInfo provides more readable information regarding rotator cuff injury than ChatGPT.美国矫形外科医师学会的OrthoInfo提供了比ChatGPT更具可读性的关于肩袖损伤的信息。
J ISAKOS. 2025 Feb 12;12:100841. doi: 10.1016/j.jisako.2025.100841.
7
Readability of patient education materials in ophthalmology: a single-institution study and systematic review.眼科患者教育材料的可读性:一项单机构研究及系统评价
BMC Ophthalmol. 2016 Aug 3;16:133. doi: 10.1186/s12886-016-0315-0.
8
Can artificial intelligence improve the readability of patient education information in gynecology?人工智能能否提高妇科患者教育信息的可读性?
Am J Obstet Gynecol. 2025 Jun 25. doi: 10.1016/j.ajog.2025.06.047.
9
The performance of ChatGPT-4 and Bing Chat in frequently asked questions about glaucoma.ChatGPT-4和必应聊天在青光眼常见问题方面的表现。
Eur J Ophthalmol. 2025 Jul;35(4):1323-1328. doi: 10.1177/11206721251321197. Epub 2025 Feb 19.
10
Accuracy and Readability of ChatGPT Responses to Patient-Centric Strabismus Questions.ChatGPT对以患者为中心的斜视问题的回答的准确性和可读性。
J Pediatr Ophthalmol Strabismus. 2025 May-Jun;62(3):220-227. doi: 10.3928/01913913-20250110-02. Epub 2025 Feb 19.