Suppr超能文献

评估 ChatGPT 在回答变应性鼻炎和慢性鼻-鼻窦炎相关问题方面的表现。

Evaluating ChatGPT's Performance in Answering Questions About Allergic Rhinitis and Chronic Rhinosinusitis.

机构信息

Department of Otolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-Sen University, Guangzhou, China.

Department of Allergy, The Third Affiliated Hospital of Sun Yat-Sen University, Guangzhou, China.

出版信息

Otolaryngol Head Neck Surg. 2024 Aug;171(2):571-577. doi: 10.1002/ohn.832. Epub 2024 May 26.

Abstract

OBJECTIVE

This study aims to evaluate the accuracy of ChatGPT in answering allergic rhinitis (AR) and chronic rhinosinusitis (CRS) related questions.

STUDY DESIGN

This is a cross-sectional study.

SETTING

Each question was inputted as a separate, independent prompt.

METHODS

Responses to AR (n = 189) and CRS (n = 242) related questions, generated by GPT-3.5 and GPT-4, were independently graded for accuracy by 2 senior rhinology professors, with disagreements adjudicated by a third reviewer.

RESULTS

Overall, ChatGPT demonstrated a satisfactory performance, accurately answering over 80% of questions across all categories. Specifically, GPT-4.0's accuracy in responding to AR-related questions significantly exceeded that of GPT-3.5, but distinction not evident in CRS-related questions. Patient-originated questions had a significantly higher accuracy compared to doctor-originated questions when utilizing GPT-4.0 to respond to AR-related questions. This discrepancy was not observed with GPT-3.5 or in the context of CRS-related questions. Across different types of content, ChatGPT excelled in covering basic knowledge, prevention, and emotion for AR and CRS. However, it experienced challenges when addressing questions about recent advancements, a trend consistent across both GPT-3.5 and GPT-4.0 iterations. Importantly, the accuracy of responses remained unaffected when questions were posed in Chinese.

CONCLUSION

Our findings suggest ChatGPT's capability to convey accurate information for AR and CRS patients, and offer insights into its performance across various domains, guiding its utilization and improvement.

摘要

目的

本研究旨在评估 ChatGPT 在回答变应性鼻炎(AR)和慢性鼻-鼻窦炎(CRS)相关问题方面的准确性。

设计

这是一项横断面研究。

设置

每个问题都作为一个单独的、独立的提示输入。

方法

由 2 位资深鼻科学教授独立评估 GPT-3.5 和 GPT-4 生成的 AR(n=189)和 CRS(n=242)相关问题的回答的准确性,有分歧的由第 3 位审稿人裁决。

结果

总体而言,ChatGPT 表现出令人满意的性能,在所有类别中准确回答了超过 80%的问题。具体而言,GPT-4.0 回答 AR 相关问题的准确性明显高于 GPT-3.5,但在 CRS 相关问题中则不明显。当利用 GPT-4.0 回答 AR 相关问题时,患者提出的问题比医生提出的问题具有更高的准确性。在 GPT-3.5 或 CRS 相关问题中则没有观察到这种差异。在不同类型的内容中,ChatGPT 在涵盖 AR 和 CRS 的基础知识、预防和情感方面表现出色。然而,在涉及最近进展的问题时,它遇到了挑战,这种趋势在 GPT-3.5 和 GPT-4.0 迭代中都存在。重要的是,当以中文提问时,回答的准确性不受影响。

结论

我们的研究结果表明,ChatGPT 有能力为 AR 和 CRS 患者传达准确的信息,并深入了解其在各个领域的表现,指导其利用和改进。

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验