Gaudiani Michael A, Castle Joshua P, Abbas Muhammad J, Pratt Brittaney A, Myles Marquisha D, Moutzouros Vasilios, Lynch T Sean
Department of Orthopedic Surgery, Henry Ford Health, Detroit, Michigan, U.S.A.
Michigan State University College of Human Medicine, Detroit, Michigan, U.S.A.
Arthrosc Sports Med Rehabil. 2024 Apr 9;6(3):100939. doi: 10.1016/j.asmr.2024.100939. eCollection 2024 Jun.
To replicate a patient's internet search to evaluate ChatGPT's appropriateness in answering common patient questions about anterior cruciate ligament reconstruction compared with a Google web search.
A Google web search was performed by searching the term "anterior cruciate ligament reconstruction." The top 20 frequently asked questions and responses were recorded. The prompt "What are the 20 most popular patient questions related to 'anterior cruciate ligament reconstruction?'" was input into ChatGPT and questions and responses were recorded. Questions were classified based on the Rothwell system and responses assessed via Flesch-Kincaid Grade Level, correctness, and completeness were for both Google web search and ChatGPT.
Three of 20 (15%) questions were similar between Google web search and ChatGPT. The most common question types among the Google web search were value (8/20, 40%), fact (7/20, 35%), and policy (5/20, 25%). The most common question types amongst the ChatGPT search were fact (12/20, 60%), policy (6/20, 30%), and value (2/20, 10%). Mean Flesch-Kincaid Grade Level for Google web search responses was significantly lower (11.8 ± 3.8 vs 14.3 ± 2.2; = .003) than for ChatGPT responses. The mean correctness for Google web search question answers was 1.47 ± 0.5, and mean completeness was 1.36 ± 0.5. Mean correctness for ChatGPT answers was 1.8 ± 0.4 and mean completeness was 1.9 ± 0.3, which were both significantly greater than Google web search answers ( = .03 and = .0003).
ChatGPT-4 generated more accurate and complete responses to common patient questions about anterior cruciate ligament reconstruction than Google's search engine.
The use of artificial intelligence such as ChatGPT is expanding. It is important to understand the quality of information as well as how the results of ChatGPT queries compare with those from Google web searches.
复制患者的网络搜索,以评估ChatGPT在回答有关前交叉韧带重建的常见患者问题方面与谷歌网络搜索相比的适用性。
通过搜索术语“前交叉韧带重建”进行谷歌网络搜索。记录前20个常见问题及答案。将提示语“与‘前交叉韧带重建’相关的20个最常见患者问题是什么?”输入ChatGPT,并记录问题及答案。根据罗斯韦尔系统对问题进行分类,并通过弗莱什-金凯德年级水平、正确性和完整性对谷歌网络搜索和ChatGPT的答案进行评估。
谷歌网络搜索和ChatGPT之间的20个问题中有3个(15%)相似。谷歌网络搜索中最常见的问题类型是价值类(8/20,40%)、事实类(7/20,35%)和政策类(5/20,25%)。ChatGPT搜索中最常见的问题类型是事实类(12/20,60%)、政策类(6/20,30%)和价值类(2/20,10%)。谷歌网络搜索答案的平均弗莱什-金凯德年级水平显著低于ChatGPT答案(11.8±3.8对14.3±2.2;P = 0.003)。谷歌网络搜索问题答案的平均正确性为1.47±0.5,平均完整性为1.36±0.5。ChatGPT答案的平均正确性为1.8±0.4,平均完整性为1.9±0.3,均显著高于谷歌网络搜索答案(P = 0.03和P = 0.0003)。
ChatGPT-4对有关前交叉韧带重建的常见患者问题生成的回答比谷歌搜索引擎更准确、更完整。
ChatGPT等人工智能的应用正在扩大。了解信息质量以及ChatGPT查询结果与谷歌网络搜索结果的比较情况非常重要。