• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

大语言模型是解决患者关于拇外翻常见担忧的有用资源吗?一项可读性分析。

Are large language models a useful resource to address common patient concerns on hallux valgus? A readability analysis.

作者信息

Hlavinka William J, Sontam Tarun R, Gupta Anuj, Croen Brett J, Abdullah Mohammed S, Humbyrd Casey J

机构信息

Texas A&M School of Medicine, Baylor University Medical Center, Department of Medical Education, 3500 Gaston Avenue, 6-Roberts, Dallas, TX 75246, USA.

Department of Orthopedic Surgery, University of Pennsylvania Health System, 51 N 39th St, Philadelphia, PA 19104, USA.

出版信息

Foot Ankle Surg. 2025 Jan;31(1):15-19. doi: 10.1016/j.fas.2024.08.002. Epub 2024 Aug 6.

DOI:10.1016/j.fas.2024.08.002
PMID:39117535
Abstract

BACKGROUND

This study evaluates the accuracy and readability of Google, ChatGPT-3.5, and 4.0 (two versions of an artificial intelligence model) responses to common questions regarding bunion surgery.

METHODS

A Google search of "bunionectomy" was performed, and the first ten questions under "People Also Ask" were recorded. ChatGPT-3.5 and 4.0 were asked these ten questions individually, and their answers were analyzed using the Flesch-Kincaid Reading Ease and Gunning-Fog Level algorithms.

RESULTS

When compared to Google, ChatGPT-3.5 and 4.0 had a larger word count with 315 ± 39 words (p < .0001) and 294 ± 39 words (p < .0001), respectively. A significant difference was found between ChatGPT-3.5 and 4.0 compared to Google using Flesch-Kincaid Reading Ease (p < .0001).

CONCLUSIONS

Our findings demonstrate that ChatGPT provided significantly lengthier responses than Google and there was a significant difference in reading ease. Both platforms exceeded the seventh to eighth-grade reading level of the U.S.

LEVEL OF EVIDENCE

N/A.

摘要

背景

本研究评估了谷歌、ChatGPT-3.5和4.0(人工智能模型的两个版本)对有关拇囊炎手术常见问题的回答的准确性和可读性。

方法

在谷歌上搜索“拇囊切除术”,记录“相关问题”下的前十个问题。分别向ChatGPT-3.5和4.0提出这十个问题,并使用弗莱什-金凯德易读性算法和冈宁-福格指数算法分析它们的答案。

结果

与谷歌相比,ChatGPT-3.5和4.0的单词数更多,分别为315±39个单词(p<.0001)和294±39个单词(p<.0001)。使用弗莱什-金凯德易读性算法发现,与谷歌相比,ChatGPT-3.5和4.0之间存在显著差异(p<.0001)。

结论

我们的研究结果表明,ChatGPT提供的回答比谷歌长得多,并且在易读性方面存在显著差异。两个平台都超过了美国七年级到八年级的阅读水平。

证据水平

无。

相似文献

1
Are large language models a useful resource to address common patient concerns on hallux valgus? A readability analysis.大语言模型是解决患者关于拇外翻常见担忧的有用资源吗?一项可读性分析。
Foot Ankle Surg. 2025 Jan;31(1):15-19. doi: 10.1016/j.fas.2024.08.002. Epub 2024 Aug 6.
2
Appropriateness and readability of Google Bard and ChatGPT-3.5 generated responses for surgical treatment of glaucoma.谷歌巴德和 ChatGPT-3.5 生成的青光眼手术治疗回复的适宜性和可读性。
Rom J Ophthalmol. 2024 Jul-Sep;68(3):243-248. doi: 10.22336/rjo.2024.45.
3
Is Information About Musculoskeletal Malignancies From Large Language Models or Web Resources at a Suitable Reading Level for Patients?来自大语言模型或网络资源的关于肌肉骨骼恶性肿瘤的信息对患者来说是否处于合适的阅读水平?
Clin Orthop Relat Res. 2025 Feb 1;483(2):306-315. doi: 10.1097/CORR.0000000000003263. Epub 2024 Sep 25.
4
Dr. Google vs. Dr. ChatGPT: Exploring the Use of Artificial Intelligence in Ophthalmology by Comparing the Accuracy, Safety, and Readability of Responses to Frequently Asked Patient Questions Regarding Cataracts and Cataract Surgery.谷歌医生与ChatGPT医生:通过比较关于白内障及白内障手术的常见患者问题的回答的准确性、安全性和可读性,探索人工智能在眼科领域的应用。
Semin Ophthalmol. 2024 Aug;39(6):472-479. doi: 10.1080/08820538.2024.2326058. Epub 2024 Mar 22.
5
Generative artificial intelligence chatbots may provide appropriate informational responses to common vascular surgery questions by patients.生成式人工智能聊天机器人可能会为患者关于常见血管外科问题提供恰当的信息性回复。
Vascular. 2025 Feb;33(1):229-237. doi: 10.1177/17085381241240550. Epub 2024 Mar 18.
6
The impact of internet resources and artificial intelligence on information on myringotomy tubes.互联网资源和人工智能对鼓膜切开术管相关信息的影响
Eur Arch Otorhinolaryngol. 2025 Apr;282(4):2149-2153. doi: 10.1007/s00405-024-09148-0. Epub 2024 Dec 12.
7
American academy of Orthopedic Surgeons' OrthoInfo provides more readable information regarding meniscus injury than ChatGPT-4 while information accuracy is comparable.美国矫形外科医师学会的OrthoInfo在半月板损伤方面提供了比ChatGPT-4更具可读性的信息,而信息准确性相当。
J ISAKOS. 2025 Apr;11:100843. doi: 10.1016/j.jisako.2025.100843. Epub 2025 Feb 21.
8
Readability, reliability and quality of responses generated by ChatGPT, gemini, and perplexity for the most frequently asked questions about pain.ChatGPT、Gemini和Perplexity针对最常见疼痛问题生成的回答的可读性、可靠性和质量。
Medicine (Baltimore). 2025 Mar 14;104(11):e41780. doi: 10.1097/MD.0000000000041780.
9
Information Quality and Readability: ChatGPT's Responses to the Most Common Questions About Spinal Cord Injury.信息质量与可读性:ChatGPT 对脊髓损伤常见问题的回答
World Neurosurg. 2024 Jan;181:e1138-e1144. doi: 10.1016/j.wneu.2023.11.062. Epub 2023 Nov 22.
10
Readability analysis of ChatGPT's responses on lung cancer.肺癌相关问题的 ChatGPT 回复可读性分析。
Sci Rep. 2024 Jul 26;14(1):17234. doi: 10.1038/s41598-024-67293-2.

引用本文的文献

1
Is it a pediatric orthopaedic urgency or not? Can ChatGPT answer this question?这是否属于小儿骨科急症?ChatGPT能回答这个问题吗?
J Orthop Surg Res. 2025 Jun 4;20(1):567. doi: 10.1186/s13018-025-05981-z.
2
Evolution of patient education materials from large-language artificial intelligence models on complex regional pain syndrome: are patients learning?基于大语言人工智能模型的复杂区域疼痛综合征患者教育材料的演变:患者有在学习吗?
Proc (Bayl Univ Med Cent). 2025 Feb 28;38(3):221-226. doi: 10.1080/08998280.2025.2470033. eCollection 2025.