• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

探究大语言模型在屈光手术相关问题上的作用。

Investigating the role of large language models on questions about refractive surgery.

作者信息

Demir Suleyman

机构信息

Adana 5 Ocak State Hospital, Department of Ophthalmology, Adana, Turkey.

出版信息

Int J Med Inform. 2025 Mar;195:105787. doi: 10.1016/j.ijmedinf.2025.105787. Epub 2025 Jan 6.

DOI:10.1016/j.ijmedinf.2025.105787
PMID:39787660
Abstract

BACKGROUND

Large language models (LLMs) are becoming increasingly popular and are playing an important role in providing accurate clinical information to both patients and physicians. This study aimed to investigate the effectiveness of ChatGPT-4.0, Google Gemini, and Microsoft Copilot LLMs for responding to patient questions regarding refractive surgery.

METHODS

The LLMs' responses to 25 questions about refractive surgery, which are frequently asked by patients, were evaluated by two ophthalmologists using a 5-point Likert scale, with scores ranging from 1 to 5. Furthermore, the DISCERN scale was used to assess the reliability of the language models' responses, whereas the Flesch Reading Ease and Flesch-Kincaid Grade Level indices were used to evaluate readability.

RESULTS

Significant differences were found among all three LLMs in the Likert scores (p = 0.022). Pairwise comparisons revealed that ChatGPT-4.0's Likert score was significantly higher than that of Microsoft Copilot, while no significant difference was found when compared to Google Gemini (p = 0.005 and p = 0.087, respectively). In terms of reliability, ChatGPT-4.0 stood out, receiving the highest DISCERN scores among the three LLMs. However, in terms of readability, ChatGPT-4.0 received the lowest score.

CONCLUSIONS

ChatGPT-4.0's responses to inquiries regarding refractive surgery were more intricate for patients compared to other language models; however, the information provided was more dependable and accurate.

摘要

背景

大语言模型(LLMs)越来越受欢迎,在为患者和医生提供准确的临床信息方面发挥着重要作用。本研究旨在调查ChatGPT-4.0、谷歌Gemini和微软Copilot大语言模型在回答患者关于屈光手术问题方面的有效性。

方法

两名眼科医生使用5分制李克特量表(分数范围为1至5)对大语言模型对患者经常问到的25个关于屈光手术问题的回答进行评估。此外,使用DISCERN量表评估语言模型回答的可靠性,而使用弗莱什易读性和弗莱什-金凯德年级水平指数评估可读性。

结果

所有三个大语言模型的李克特得分存在显著差异(p = 0.022)。两两比较显示,ChatGPT-4.0的李克特得分显著高于微软Copilot,而与谷歌Gemini相比未发现显著差异(分别为p = 0.005和p = 0.087)。在可靠性方面,ChatGPT-4.0表现突出,在三个大语言模型中获得最高的DISCERN分数。然而,在可读性方面,ChatGPT-4.0得分最低。

结论

与其他语言模型相比,ChatGPT-4.0对屈光手术相关询问的回答对患者来说更为复杂;然而,所提供的信息更可靠、准确。

相似文献

1
Investigating the role of large language models on questions about refractive surgery.探究大语言模型在屈光手术相关问题上的作用。
Int J Med Inform. 2025 Mar;195:105787. doi: 10.1016/j.ijmedinf.2025.105787. Epub 2025 Jan 6.
2
Evaluation of Responses to Questions About Keratoconus Using ChatGPT-4.0, Google Gemini and Microsoft Copilot: A Comparative Study of Large Language Models on Keratoconus.使用ChatGPT-4.0、谷歌Gemini和微软Copilot评估圆锥角膜相关问题的回答:大型语言模型在圆锥角膜方面的比较研究
Eye Contact Lens. 2025 Mar 1;51(3):e107-e111. doi: 10.1097/ICL.0000000000001158. Epub 2024 Dec 4.
3
Evaluation of the reliability and readability of answers given by chatbots to frequently asked questions about endophthalmitis: A cross-sectional study on chatbots.评估聊天机器人对眼内炎常见问题回答的可靠性和可读性:一项关于聊天机器人的横断面研究。
Health Informatics J. 2024 Oct-Dec;30(4):14604582241304679. doi: 10.1177/14604582241304679.
4
Proficiency, Clarity, and Objectivity of Large Language Models Versus Specialists' Knowledge on COVID-19's Impacts in Pregnancy: Cross-Sectional Pilot Study.大型语言模型在新冠肺炎对妊娠影响方面的熟练度、清晰度和客观性与专家知识对比:横断面试点研究
JMIR Form Res. 2025 Feb 5;9:e56126. doi: 10.2196/56126.
5
Assessing the Responses of Large Language Models (ChatGPT-4, Claude 3, Gemini, and Microsoft Copilot) to Frequently Asked Questions in Retinopathy of Prematurity: A Study on Readability and Appropriateness.评估大型语言模型(ChatGPT-4、Claude 3、Gemini和Microsoft Copilot)对早产儿视网膜病变常见问题的回答:一项关于可读性和适宜性的研究
J Pediatr Ophthalmol Strabismus. 2025 Mar-Apr;62(2):84-95. doi: 10.3928/01913913-20240911-05. Epub 2024 Oct 28.
6
Assessing the Responses of Large Language Models (ChatGPT-4, Gemini, and Microsoft Copilot) to Frequently Asked Questions in Breast Imaging: A Study on Readability and Accuracy.评估大语言模型(ChatGPT-4、Gemini和Microsoft Copilot)对乳腺成像常见问题的回答:可读性和准确性研究
Cureus. 2024 May 9;16(5):e59960. doi: 10.7759/cureus.59960. eCollection 2024 May.
7
Evaluating the reliability of the responses of large language models to keratoconus-related questions.评估大语言模型对圆锥角膜相关问题回答的可靠性。
Clin Exp Optom. 2024 Oct 24:1-8. doi: 10.1080/08164622.2024.2419524.
8
Comparison of ChatGPT-4o, Google Gemini 1.5 Pro, Microsoft Copilot Pro, and Ophthalmologists in the management of uveitis and ocular inflammation: A comparative study of large language models.ChatGPT-4o、谷歌Gemini 1.5 Pro、微软Copilot Pro与眼科医生在葡萄膜炎和眼部炎症管理中的比较:大型语言模型的对比研究
J Fr Ophtalmol. 2025 Apr;48(4):104468. doi: 10.1016/j.jfo.2025.104468. Epub 2025 Mar 13.
9
Microsoft Copilot Provides More Accurate and Reliable Information About Anterior Cruciate Ligament Injury and Repair Than ChatGPT and Google Gemini; However, No Resource Was Overall the Best.与ChatGPT和谷歌Gemini相比,微软Copilot能提供关于前交叉韧带损伤与修复的更准确、更可靠的信息;然而,没有一种资源在各方面都是最佳的。
Arthrosc Sports Med Rehabil. 2024 Nov 19;7(2):101043. doi: 10.1016/j.asmr.2024.101043. eCollection 2025 Apr.
10
Parental education in pediatric dysphagia: A comparative analysis of three large language models.儿科吞咽困难中的家长教育:三种大型语言模型的比较分析
J Pediatr Gastroenterol Nutr. 2025 Jul;81(1):18-26. doi: 10.1002/jpn3.70069. Epub 2025 May 8.

引用本文的文献

1
Evaluating artificial intelligence chatbots' responses to gynecomastia inquiries: Comparative study of information quality, readability, and guideline consistency.评估人工智能聊天机器人对男性乳房发育症咨询的回复:信息质量、可读性和指南一致性的比较研究
Digit Health. 2025 Aug 26;11:20552076251367645. doi: 10.1177/20552076251367645. eCollection 2025 Jan-Dec.