Suppr超能文献

ChatGPT和谷歌对与肩袖修复相关的最常见患者问题大多提供了极佳或令人满意的回答。

ChatGPT and Google Provide Mostly Excellent or Satisfactory Responses to the Most Frequently Asked Patient Questions Related to Rotator Cuff Repair.

作者信息

Megalla Martinus, Hahn Alexander K, Bauer Jordan A, Windsor Jordan T, Grace Zachary T, Gedman Marissa A, Arciero Robert A

机构信息

University of Connecticut, Farmington, Connecticut, U.S.A.

Hackensack Meridian School of Medicine, Nutley, New Jersey, U.S.A.

出版信息

Arthrosc Sports Med Rehabil. 2024 Jun 25;6(5):100963. doi: 10.1016/j.asmr.2024.100963. eCollection 2024 Oct.

Abstract

PURPOSE

To assess the differences in frequently asked questions (FAQs) and responses related to rotator cuff surgery between Google and ChatGPT.

METHODS

Both Google and ChatGPT (version 3.5) were queried for the top 10 FAQs using the search term "rotator cuff repair." Questions were categorized according to Rothwell's classification. In addition to questions and answers for each website, the source that the answer was pulled from was noted and assigned a category (academic, medical practice, etc). Responses were also graded as "excellent response not requiring clarification" (1), "satisfactory requiring minimal clarification" (2), "satisfactory requiring moderate clarification" (3), or "unsatisfactory requiring substantial clarification" (4).

RESULTS

Overall, 30% of questions were similar between what Google and ChatGPT deemed to be the most FAQs. For questions from Google web search, most answers came from medical practices (40%). For ChatGPT, most answers were provided by academic sources (90%). For numerical questions, ChatGPT and Google provided similar responses for 30% of questions. For most of the questions, both Google and ChatGPT responses were either "excellent" or "satisfactory requiring minimal clarification." Google had 1 response rated as satisfactory requiring moderate clarification, whereas ChatGPT had 2 responses rated as unsatisfactory.

CONCLUSIONS

Both Google and ChatGPT offer mostly excellent or satisfactory responses to the most FAQs regarding rotator cuff repair. However, ChatGPT may provide inaccurate or even fabricated answers and associated citations.

CLINICAL RELEVANCE

In general, the quality of online medical content is low. As artificial intelligence develops and becomes more widely used, it is important to assess the quality of the information patients are receiving from this technology.

摘要

目的

评估谷歌和ChatGPT之间与肩袖手术相关的常见问题(FAQ)及回答的差异。

方法

使用搜索词“肩袖修复”查询谷歌和ChatGPT(3.5版本)的前10个常见问题。问题根据罗斯韦尔分类法进行分类。除了每个网站的问题和答案外,还记录了答案的来源并分配了一个类别(学术、医疗实践等)。回答也被评为“无需澄清的优秀回答”(1)、“需要最少澄清的满意回答”(2)、“需要适度澄清的满意回答”(3)或“需要大量澄清的不满意回答”(4)。

结果

总体而言,谷歌和ChatGPT认为的最常见问题中有30%是相似的。对于谷歌网络搜索的问题,大多数答案来自医疗实践(40%)。对于ChatGPT,大多数答案由学术来源提供(90%)。对于数值问题,ChatGPT和谷歌对30%的问题提供了相似的回答。对于大多数问题,谷歌和ChatGPT的回答要么是“优秀”,要么是“需要最少澄清的满意回答”。谷歌有1个回答被评为需要适度澄清的满意回答,而ChatGPT有2个回答被评为不满意回答。

结论

谷歌和ChatGPT对于肩袖修复的大多数常见问题大多提供了优秀或满意的回答。然而,ChatGPT可能会提供不准确甚至编造的答案及相关引用。

临床意义

一般来说,在线医疗内容质量较低。随着人工智能的发展和更广泛的应用,评估患者从这项技术获得的信息质量很重要。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dfee/11551354/cb189879a7bf/gr1.jpg

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验