Suppr超能文献

针对泌尿外科患者的ChatGPT输出信息质量及适用性

Quality of information and appropriateness of ChatGPT outputs for urology patients.

作者信息

Cocci Andrea, Pezzoli Marta, Lo Re Mattia, Russo Giorgio Ivan, Asmundo Maria Giovanna, Fode Mikkel, Cacciamani Giovanni, Cimino Sebastiano, Minervini Andrea, Durukan Emil

机构信息

Urology Section, University of Florence, Florence, Italy.

Urology Section, University of Catania, Catania, Italy.

出版信息

Prostate Cancer Prostatic Dis. 2024 Mar;27(1):103-108. doi: 10.1038/s41391-023-00705-y. Epub 2023 Jul 29.

Abstract

BACKGROUND

The proportion of health-related searches on the internet is continuously growing. ChatGPT, a natural language processing (NLP) tool created by OpenAI, has been gaining increasing user attention and can potentially be used as a source for obtaining information related to health concerns. This study aims to analyze the quality and appropriateness of ChatGPT's responses to Urology case studies compared to those of a urologist.

METHODS

Data from 100 patient case studies, comprising patient demographics, medical history, and urologic complaints, were sequentially inputted into ChatGPT, one by one. A question was posed to determine the most likely diagnosis, suggested examinations, and treatment options. The responses generated by ChatGPT were then compared to those provided by a board-certified urologist who was blinded to ChatGPT's responses and graded on a 5-point Likert scale based on accuracy, comprehensiveness, and clarity as criterias for appropriateness. The quality of information was graded based on the section 2 of the DISCERN tool and readability assessments were performed using the Flesch Reading Ease (FRE) and Flesch-Kincaid Reading Grade Level (FKGL) formulas.

RESULTS

52% of all responses were deemed appropriate. ChatGPT provided more appropriate responses for non-oncology conditions (58.5%) compared to oncology (52.6%) and emergency urology cases (11.1%) (p = 0.03). The median score of the DISCERN tool was 15 (IQR = 5.3) corresponding to a quality score of poor. The ChatGPT responses demonstrated a college graduate reading level, as indicated by the median FRE score of 18 (IQR = 21) and the median FKGL score of 15.8 (IQR = 3).

CONCLUSIONS

ChatGPT serves as an interactive tool for providing medical information online, offering the possibility of enhancing health outcomes and patient satisfaction. Nevertheless, the insufficient appropriateness and poor quality of the responses on Urology cases emphasizes the importance of thorough evaluation and use of NLP-generated outputs when addressing health-related concerns.

摘要

背景

互联网上与健康相关的搜索比例在持续增长。ChatGPT是OpenAI创建的一种自然语言处理(NLP)工具,越来越受到用户关注,并有可能被用作获取与健康问题相关信息的来源。本研究旨在分析ChatGPT对泌尿外科病例研究的回答与泌尿外科医生回答相比的质量和适当性。

方法

将100个患者病例研究的数据,包括患者人口统计学、病史和泌尿外科主诉,逐一依次输入ChatGPT。提出一个问题以确定最可能的诊断、建议的检查和治疗方案。然后将ChatGPT生成的回答与一位对ChatGPT的回答不知情的获得委员会认证的泌尿外科医生提供的回答进行比较,并根据准确性、全面性和清晰度作为适当性标准,采用5点李克特量表进行评分。信息质量根据DISCERN工具的第2节进行评分,并使用弗莱什易读性(FRE)和弗莱什-金凯德阅读年级水平(FKGL)公式进行可读性评估。

结果

所有回答中有52%被认为是适当的。与肿瘤学(52.6%)和泌尿外科急诊病例(11.1%)相比,ChatGPT对非肿瘤疾病提供了更适当的回答(58.5%)(p = 0.03)。DISCERN工具的中位数分数为15(四分位距 = 5.3),对应质量评分为差。ChatGPT的回答显示出大学毕业生的阅读水平,FRE中位数分数为18(四分位距 = 21),FKGL中位数分数为15.8(四分位距 = 3)表明了这一点。

结论

ChatGPT作为一种在线提供医疗信息的交互式工具,为改善健康结果和患者满意度提供了可能性。然而,泌尿外科病例回答的适当性不足和质量较差,强调了在处理与健康相关问题时,对NLP生成的输出进行全面评估和使用的重要性。

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验