Suppr超能文献

ChatGPT 在不同耳鼻喉科亚专业中的测验技能:对 2576 道选择题和多选题进行 board certification 准备的分析。

ChatGPT's quiz skills in different otolaryngology subspecialties: an analysis of 2576 single-choice and multiple-choice board certification preparation questions.

机构信息

Department of Otolaryngology, Head and Neck Surgery, School of Medicine, Technical University of Munich (TUM), Ismaningerstrasse 22, 81675, Munich, Germany.

Department of Otorhinolaryngology, Head and Neck Surgery, Medical Faculty, University of Cologne, 50937, Cologne, Germany.

出版信息

Eur Arch Otorhinolaryngol. 2023 Sep;280(9):4271-4278. doi: 10.1007/s00405-023-08051-4. Epub 2023 Jun 7.

Abstract

PURPOSE

With the increasing adoption of artificial intelligence (AI) in various domains, including healthcare, there is growing acceptance and interest in consulting AI models to provide medical information and advice. This study aimed to evaluate the accuracy of ChatGPT's responses to practice quiz questions designed for otolaryngology board certification and decipher potential performance disparities across different otolaryngology subspecialties.

METHODS

A dataset covering 15 otolaryngology subspecialties was collected from an online learning platform funded by the German Society of Oto-Rhino-Laryngology, Head and Neck Surgery, designed for board certification examination preparation. These questions were entered into ChatGPT, with its responses being analyzed for accuracy and variance in performance.

RESULTS

The dataset included 2576 questions (479 multiple-choice and 2097 single-choice), of which 57% (n = 1475) were answered correctly by ChatGPT. An in-depth analysis of question style revealed that single-choice questions were associated with a significantly higher rate (p < 0.001) of correct responses (n = 1313; 63%) compared to multiple-choice questions (n = 162; 34%). Stratified by question categories, ChatGPT yielded the highest rate of correct responses (n = 151; 72%) in the field of allergology, whereas 7 out of 10 questions (n = 65; 71%) on legal otolaryngology aspects were answered incorrectly.

CONCLUSION

The study reveals ChatGPT's potential as a supplementary tool for otolaryngology board certification preparation. However, its propensity for errors in certain otolaryngology areas calls for further refinement. Future research should address these limitations to improve ChatGPT's educational use. An approach, with expert collaboration, is recommended for the reliable and accurate integration of such AI models.

摘要

目的

随着人工智能(AI)在各个领域的广泛应用,包括医疗保健领域,人们越来越接受并热衷于咨询 AI 模型以提供医学信息和建议。本研究旨在评估 ChatGPT 对耳鼻喉科委员会认证实践测验问题的回答的准确性,并揭示不同耳鼻喉科亚专业之间潜在的表现差异。

方法

从一个由德国耳鼻喉科协会资助的在线学习平台上收集了涵盖 15 个耳鼻喉科亚专业的数据集,该数据集专为委员会认证考试准备而设计。将这些问题输入到 ChatGPT 中,分析其回答的准确性和表现差异。

结果

该数据集包括 2576 个问题(479 个多项选择和 2097 个单项选择),其中 57%(n=1475)被 ChatGPT 回答正确。对问题风格的深入分析表明,单项选择问题与更高的正确回答率(p<0.001)显著相关(n=1313;63%),相比之下,多项选择问题(n=162;34%)的正确回答率较低。按问题类别分层,ChatGPT 在变态反应学领域产生了最高的正确回答率(n=151;72%),而在法律耳鼻喉科方面的 10 个问题中有 7 个(n=65;71%)回答错误。

结论

本研究揭示了 ChatGPT 在耳鼻喉科委员会认证准备中的潜在作用。然而,它在某些耳鼻喉科领域的错误倾向需要进一步改进。未来的研究应该解决这些限制,以提高 ChatGPT 的教育用途。建议采用专家合作的方法,以可靠和准确地整合此类 AI 模型。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1690/10382366/15ddde19c19a/405_2023_8051_Fig1_HTML.jpg

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验