Department of Urology, Sidney Kimmel Medical College, Thomas Jefferson University, Philadelphia, Pennsylvania, USA.
Can J Urol. 2024 Oct;31(5):12013-12018.
Patients seek support from online resources when facing a troubling urologic cancer diagnosis. Physician-written resources exceed the recommended 6-8th grade reading level, creating confusion and driving patients towards unregulated online materials like AI chatbots. We aim to compare the readability and quality of patient education on ChatGPT against Epic and Urology Care Foundation (UCF).
We analyzed prostate, bladder, and kidney cancer content from ChatGPT, Epic, and UCF. We further studied readability-adjusted responses using specific AI prompting (ChatGPT-a) and Epic material designated as Easy to Read. Blinded reviewers completed descriptive textual analysis, readability analysis via six validated formulas, and quality analysis via DISCERN, PEMAT, and Likert tools.
Epic met the recommended grade level, while UCF and ChatGPT exceeded it (5.81 vs. 8.44 vs. 12.16, p < 0.001). ChatGPT text was longer with more complex wording (p < 0.001). Quality was fair for Epic, good for UCF, and excellent for ChatGPT (49.5 vs. 61.67 vs. 64.33). Actionability was overall poor but particularly lowest (37%) for Epic. On qualitative analysis, Epic lagged on all quality measures. When adjusted for user education level (ChatGPT-a and Epic Easy to Read), readability improved (7.50 and 3.53), but only ChatGPT-a retained high quality.
Online urologic oncology patient materials largely exceed the average American's literacy level and often lack real-world utility for patients. Our ChatGPT-a model indicates that AI technology can improve accessibility and usefulness. With development, a healthcare-specific AI program may help providers create content that is accessible and personalized to improve shared decision-making for urology patients.
当面临困扰性的泌尿科癌症诊断时,患者会从在线资源中寻求支持。医生撰写的资源超过了推荐的 6-8 年级阅读水平,这造成了混淆,并促使患者转向不受监管的在线材料,如 AI 聊天机器人。我们旨在比较 ChatGPT 与 Epic 和 Urology Care Foundation(UCF)在患者教育方面的可读性和质量。
我们分析了 ChatGPT、Epic 和 UCF 中的前列腺癌、膀胱癌和肾癌内容。我们进一步研究了使用特定 AI 提示(ChatGPT-a)和 Epic 中指定为“易于阅读”的材料进行可读性调整后的回复。盲审员完成了描述性文本分析、使用六个经过验证的公式进行可读性分析以及使用 DISCERN、PEMAT 和 Likert 工具进行质量分析。
Epic 符合推荐的年级水平,而 UCF 和 ChatGPT 则超过了该水平(5.81 对 8.44 对 12.16,p<0.001)。ChatGPT 文本更长,用词更复杂(p<0.001)。Epic 的质量为良好,UCF 的质量为优秀,而 ChatGPT 的质量为极好(49.5 对 61.67 对 64.33)。行动性总体较差,但 Epic 尤其低(37%)。在定性分析中,Epic 在所有质量指标上都滞后。当根据用户教育水平调整(ChatGPT-a 和 Epic Easy to Read)时,可读性有所提高(7.50 和 3.53),但只有 ChatGPT-a 保持了高质量。
在线泌尿科肿瘤患者材料在很大程度上超过了普通美国人的读写水平,并且往往对患者缺乏实际用途。我们的 ChatGPT-a 模型表明,人工智能技术可以提高可及性和实用性。随着发展,专门针对医疗保健的 AI 程序可能有助于提供者创建对患者可访问且个性化的内容,以改善泌尿科患者的共同决策。