• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

评估聊天机器人生成白内障手术宣传册的能力:必应人工智能、ChatGPT 3.5、ChatGPT 4、ChatSonic、谷歌巴德、Perplexity和Pi。

Assessing chatbots ability to produce leaflets on cataract surgery: Bing AI, chatGPT 3.5, chatGPT 4o, ChatSonic, Google Bard, Perplexity, and Pi.

作者信息

Thompson Polly, Thornton Richard, Ramsden Conor M

机构信息

From the The West of England Eye Unit, Royal Devon University Hospital NHS Foundation Trust, Exeter, United Kingdom (Thompson, Ramsden); Royal Eye Infirmary, Derriford Hospital, University Hospitals Plymouth NHS Foundation Trust, Plymouth, United Kingdom (Thornton); The University of Exeter, Exeter, United Kingdom (Ramsden).

出版信息

J Cataract Refract Surg. 2025 May 1;51(5):371-375. doi: 10.1097/j.jcrs.0000000000001622.

DOI:10.1097/j.jcrs.0000000000001622
PMID:39885649
Abstract

PURPOSE

To evaluate leaflets on cataract surgery produced by 7 common free chatbots.

SETTING

UK-based ophthalmologists carrying out online research.

DESIGN

Data were collected from the responses of 7 freely available online chatbots.

METHODS

Analysis of answers given by 7 chatbots (Bing AI, chatGPT 3.5, chatGPT 4o, ChatSonic, Google Bard, Perplexity, and Pi) was prompted to "make a patient information leaflet on cataract surgery." Answers were evaluated using the DISCERN instrument, Patient Education Materials Assessment Tool (PEMAT), presence of misinformation, the Flesch-Kincaid Grade level readability score, and material reliability.

RESULTS

The highest overall scored response was from ChatSonic, followed by Bing AI and then Perplexity. The lowest scoring was ChatGPT 3.5. ChatSonic achieved the highest DISCERN and PEMAT scores, and had the highest Flesch-Kincaid Grade level. The lowest DISCERN and PEMAT scores were for Pi. Only ChatGPT 3.5 included some misinformation in its response. Bing AI, ChatSonic, and Perplexity included reliable references; the other chatbots provided no references.

CONCLUSIONS

This study demonstrates a range of answers given by chatbots creating a cataract surgery leaflet, suggesting variation in their development and reliability. ChatGPT 3.5 scored the most poorly. However, ChatSonic indicated promise in how technology may be used to assist information giving in ophthalmology.

摘要

目的

评估7个常见免费聊天机器人生成的白内障手术宣传册。

背景

开展在线研究的英国眼科医生。

设计

从7个免费在线聊天机器人的回复中收集数据。

方法

促使7个聊天机器人(必应人工智能、ChatGPT 3.5、ChatGPT 4、ChatSonic、谷歌巴德、Perplexity和Pi)回答“制作一份关于白内障手术的患者信息宣传册”。使用DISCERN工具、患者教育材料评估工具(PEMAT)、错误信息的存在情况、弗莱什-金凯德年级水平可读性得分以及材料可靠性对回答进行评估。

结果

总体得分最高的回复来自ChatSonic,其次是必应人工智能,然后是Perplexity。得分最低的是ChatGPT 3.5。ChatSonic获得了最高的DISCERN和PEMAT分数,并且弗莱什-金凯德年级水平最高。Pi的DISCERN和PEMAT分数最低。只有ChatGPT 3.5在其回复中包含一些错误信息。必应人工智能、ChatSonic和Perplexity包含可靠的参考文献;其他聊天机器人未提供参考文献。

结论

本研究展示了聊天机器人在创建白内障手术宣传册时给出的一系列答案,表明它们在开发和可靠性方面存在差异。ChatGPT 3.5得分最差。然而,ChatSonic显示出利用技术辅助眼科信息提供的潜力。

相似文献

1
Assessing chatbots ability to produce leaflets on cataract surgery: Bing AI, chatGPT 3.5, chatGPT 4o, ChatSonic, Google Bard, Perplexity, and Pi.评估聊天机器人生成白内障手术宣传册的能力:必应人工智能、ChatGPT 3.5、ChatGPT 4、ChatSonic、谷歌巴德、Perplexity和Pi。
J Cataract Refract Surg. 2025 May 1;51(5):371-375. doi: 10.1097/j.jcrs.0000000000001622.
2
Evaluating the readability, quality, and reliability of responses generated by ChatGPT, Gemini, and Perplexity on the most commonly asked questions about Ankylosing spondylitis.评估ChatGPT、Gemini和Perplexity针对强直性脊柱炎最常见问题生成的回答的可读性、质量和可靠性。
PLoS One. 2025 Jun 18;20(6):e0326351. doi: 10.1371/journal.pone.0326351. eCollection 2025.
3
Talking technology: exploring chatbots as a tool for cataract patient education.技术漫谈:探索聊天机器人作为白内障患者教育工具的作用
Clin Exp Optom. 2025 Jan;108(1):56-64. doi: 10.1080/08164622.2023.2298812. Epub 2024 Jan 9.
4
Assessing ChatGPT responses to frequently asked patient questions in reconstructive urology.评估ChatGPT对重建泌尿外科常见患者问题的回答。
Urol Pract. 2025 Feb 12:101097UPJ0000000000000792. doi: 10.1097/UPJ.0000000000000792.
5
Assessment of Artificial Intelligence Chatbot Responses to Top Searched Queries About Cancer.评估人工智能聊天机器人对癌症热门搜索查询的响应
JAMA Oncol. 2023 Oct 1;9(10):1437-1440. doi: 10.1001/jamaoncol.2023.2947.
6
The performance of ChatGPT-4 and Bing Chat in frequently asked questions about glaucoma.ChatGPT-4和必应聊天在青光眼常见问题方面的表现。
Eur J Ophthalmol. 2025 Jul;35(4):1323-1328. doi: 10.1177/11206721251321197. Epub 2025 Feb 19.
7
Accuracy and Readability of Artificial Intelligence Chatbot Responses to Vasectomy-Related Questions: Public Beware.人工智能聊天机器人对输精管切除术相关问题回答的准确性和可读性:公众需谨慎。
Cureus. 2024 Aug 28;16(8):e67996. doi: 10.7759/cureus.67996. eCollection 2024 Aug.
8
Enhancing the Readability of Online Patient Education Materials Using Large Language Models: Cross-Sectional Study.使用大语言模型提高在线患者教育材料的可读性:横断面研究。
J Med Internet Res. 2025 Jun 4;27:e69955. doi: 10.2196/69955.
9
The promising role of chatbots in keratorefractive surgery patient education.聊天机器人在角膜屈光手术患者教育中的潜在作用。
J Fr Ophtalmol. 2025 Feb;48(2):104381. doi: 10.1016/j.jfo.2024.104381. Epub 2024 Dec 13.
10
American Academy of Orthopaedic Surgeons OrthoInfo provides more readable information regarding rotator cuff injury than ChatGPT.美国矫形外科医师学会的OrthoInfo提供了比ChatGPT更具可读性的关于肩袖损伤的信息。
J ISAKOS. 2025 Feb 12;12:100841. doi: 10.1016/j.jisako.2025.100841.