Suppr超能文献

评估ChatGPT生成的面向患者的鼻科学医学信息的质量和可读性。

Evaluating the quality and readability of ChatGPT-generated patient-facing medical information in rhinology.

作者信息

Fazilat Alexander Z, Brenac Camille, Kawamoto-Duran Danae, Berry Charlotte E, Alyono Jennifer, Chang Michael T, Liu David T, Patel Zara M, Tringali Stéphane, Wan Derrick C, Fieux Maxime

机构信息

Hagey Laboratory for Pediatric Regenerative Medicine, Division of Plastic and Reconstructive Surgery, Department of Surgery, Stanford University School of Medicine, Stanford, CA, USA.

Service de chirurgie plastique reconstructrice et esthétique, Hospices Civils de Lyon, Hôpital de la Croix Rousse, Lyon, F-69004, France.

出版信息

Eur Arch Otorhinolaryngol. 2025 Apr;282(4):1911-1920. doi: 10.1007/s00405-024-09180-0. Epub 2024 Dec 26.

Abstract

PURPOSE

The artificial intelligence (AI) chatbot ChatGPT has become a major tool for generating responses in healthcare. This study assessed ChatGPT's ability to generate French preoperative patient-facing medical information (PFI) in rhinology at a comparable level to material provided by an academic source, the French Society of Otorhinolaryngology (Société Française d'Otorhinolaryngologie et Chirurgie Cervico-Faciale, SFORL).

METHODS

ChatGPT and SFORL French preoperative PFI in rhinology were compared by analyzing responses to 16 questions regarding common rhinology procedures: ethmoidectomy, sphenoidotomy, septoplasty, and endonasal dacryocystorhinostomy. Twenty rhinologists assessed the clarity, comprehensiveness, accuracy, and overall quality of the information, while 24 nonmedical individuals analyzed the clarity and overall quality. Six readability formulas were used to compare readability scores.

RESULTS

Among rhinologists, no significant difference was found between ChatGPT and SFORL regarding clarity (7.61 ± 0.36 vs. 7.53 ± 0.28; p = 0.485), comprehensiveness (7.32 ± 0.77 vs. 7.58 ± 0.50; p = 0.872), and accuracy (inaccuracies: 60% vs. 40%; p = 0.228), respectively. Non-medical individuals scored the clarity of ChatGPT significantly higher than that of the SFORL (8.16 ± 1.16 vs. 6.32 ± 1.33; p < 0.0001). The non-medical individuals chose ChatGPT as the most informative source significantly more often than rhinologists (62.8% vs. 39.7%, p < 0.001).

CONCLUSION

ChatGPT-generated French preoperative PFI in rhinology was comparable to SFORL-provided PFI regarding clarity, comprehensiveness, accuracy, readability, and overall quality. This study highlights ChatGPT's potential to increase accessibility to high quality PFI and suggests its use by physicians as a complement to academic resources written by learned societies such as the SFORL.

摘要

目的

人工智能(AI)聊天机器人ChatGPT已成为医疗保健领域生成回复的主要工具。本研究评估了ChatGPT生成法语鼻科学术前患者医疗信息(PFI)的能力,其水平与学术来源——法国耳鼻咽喉科学会(Société Française d'Otorhinolaryngologie et Chirurgie Cervico-Faciale,SFORL)提供的材料相当。

方法

通过分析对16个关于常见鼻科手术(筛窦切除术、蝶窦切开术、鼻中隔成形术和鼻内泪囊鼻腔造口术)问题的回答,比较ChatGPT和SFORL的法语鼻科学术前PFI。20位鼻科医生评估了信息的清晰度、全面性、准确性和整体质量,而24位非医学人员分析了清晰度和整体质量。使用六种可读性公式比较可读性分数。

结果

在鼻科医生中,ChatGPT和SFORL在清晰度(7.61±0.36对7.53±0.28;p = 0.485)、全面性(7.32±0.77对7.58±0.50;p = 0.872)和准确性(错误率:60%对40%;p = 0.228)方面均未发现显著差异。非医学人员对ChatGPT清晰度的评分显著高于SFORL(8.16±1.16对6.32±1.33;p < 0.0001)。非医学人员选择ChatGPT作为信息最丰富来源的频率显著高于鼻科医生(62.8%对39.7%,p < 0.001)。

结论

ChatGPT生成的法语鼻科学术前PFI在清晰度、全面性、准确性、可读性和整体质量方面与SFORL提供的PFI相当。本研究突出了ChatGPT在提高高质量PFI可及性方面的潜力,并建议医生将其用作对SFORL等学术团体撰写的学术资源的补充。

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验