• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

GPT-4生成符合当前肿瘤学指南的准确且易读的患者教育材料:一项随机评估。

GPT-4 generates accurate and readable patient education materials aligned with current oncological guidelines: A randomized assessment.

作者信息

Rodler Severin, Cei Francesco, Ganjavi Conner, Checcucci Enrico, De Backer Pieter, Rivero Belenchon Ines, Taratkin Mark, Puliatti Stefano, Veccia Alessandro, Piazza Pietro, Baekelandt Loïc, Kowalewski Karl-Friedrich, Gómez Rivas Juan, Fankhauser Christian D, Moschini Marco, Gandaglia Giorgio, Campi Riccardo, De Castro Abreu Andre, Russo Giorgio I, Cocci Andrea, Maruccia Serena, Cacciamani Giovanni E

机构信息

USC Institute of Urology and Catherine and Joseph Aresty Department of Urology, University of Southern California, Los Angeles, California, United States of America.

Artificial Intelligence Center at USC Urology, USC Institute of Urology, University of Southern California, Los Angeles, California, United States of America.

出版信息

PLoS One. 2025 Jun 4;20(6):e0324175. doi: 10.1371/journal.pone.0324175. eCollection 2025.

DOI:10.1371/journal.pone.0324175
PMID:40465696
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12136319/
Abstract

INTRODUCTION AND AIM

Guideline-based patient educational materials (PEMs) empower patients and reduce misinformation, but require frequent updates and must be adapted to the readability level of patients. The aim is to assess whether generative artificial intelligence (GenAI) can provide readable, accurate, and up-to-date PEMs that can be subsequently translated into multiple languages for broad dissemination.

STUDY DESIGN AND METHODS

The European Association of Urology (EAU) guidelines for prostate, bladder, kidney, and testicular cancer were used as the knowledge base for GPT-4 to generate PEMs. Additionally, the PEMs were translated into five commonly spoken languages within the European Union (EU). The study was conducted through a single-blinded, online randomized assessment survey. After an initial pilot assessment of the GenAI-generated PEMs, thirty-two members of the Young Academic Urologists (YAU) groups evaluated the accuracy, completeness, and clarity of the original versus GPT-generated PEMs. The translation assessment involved two native speakers from different YAU groups for each language: Dutch, French, German, Italian, and Spanish. The primary outcomes were readability, accuracy, completeness, faithfulness, and clarity. Readability was measured using Flesch Kincaid Reading Ease (FKRE), Flesch Kincaid Grade Level (FKGL), Gunning Fog (GFS) scores and Smog (SI), Coleman Liau (CLI), Automated Readability (ARI) indexes. Accuracy, completeness, faithfulness, and clarity were rated on a 5-item Likert scale.

RESULTS

The mean time to create layperson PEMs based on the latest guideline by GPT-4 was 52.1 seconds. The readability scores for the 8 original PEMs were lower than for the 8 GPT-4-generated PEMs (Mean FKRE: 43.5 vs. 70.8; p < .001). The required reading education levels were higher for original PEMs compared to GPT-4 generated PEMs (Mean FKGL: 11.6 vs. 6.1; p < .001). For all urological localized cancers, the original PEMs were not significantly different from the GPT-4 generated PEMs in accuracy, completeness, and clarity. Similarly, no differences were observed for metastatic cancers. Translations of GPT-generated PEMs were rated as faithful in 77.5% of cases and clear in 67.5% of cases.

CONCLUSIONS AND RELEVANCE

GPT-4 generated PEMs have better readability levels compared to original PEMs while maintaining similar accuracy, completeness, and clarity. The use of GenAI's information extraction and language capabilities, integrated with human oversight, can significantly reduce the workload and ensure up-to-date and accurate PEMs.

PATIENT SUMMARY

Some cancer facts made for patients can be hard to read or not in the right words for those with prostate, bladder, kidney, or testicular cancer. This study used AI to quickly make short and easy-to-read content from trusted facts. Doctors checked the AI content and found that they were just as accurate, complete, and clear as the original text made for patients. They also worked well in many languages. This AI tool can assist providers in making it easier for patients to understand their cancer and the best care they can get.

摘要

引言与目的

基于指南的患者教育材料(PEMs)能增强患者能力并减少错误信息,但需要频繁更新且必须适应患者的阅读水平。目的是评估生成式人工智能(GenAI)能否提供可读、准确且最新的PEMs,随后将其翻译成多种语言以广泛传播。

研究设计与方法

欧洲泌尿外科学会(EAU)关于前列腺癌、膀胱癌、肾癌和睾丸癌的指南被用作GPT - 4生成PEMs的知识库。此外,这些PEMs被翻译成欧盟内五种常用语言。该研究通过单盲在线随机评估调查进行。在对GenAI生成的PEMs进行初步试点评估后,32名青年学术泌尿外科医生(YAU)小组的成员评估了原始PEMs与GPT生成的PEMs在准确性、完整性和清晰度方面的差异。翻译评估为每种语言安排了两名来自不同YAU小组的母语人士:荷兰语、法语、德语、意大利语和西班牙语。主要结果包括可读性、准确性、完整性、忠实性和清晰度。可读性使用弗莱什·金凯德阅读简易度(FKRE)、弗莱什·金凯德年级水平(FKGL)、冈宁雾度(GFS)得分和烟雾指数(SI)、科尔曼·廖指数(CLI)、自动可读性指数(ARI)进行测量。准确性、完整性、忠实性和清晰度通过5项李克特量表进行评分。

结果

GPT - 4根据最新指南创建面向非专业人士的PEMs的平均时间为52.1秒。8份原始PEMs的可读性得分低于8份GPT - 4生成的PEMs(平均FKRE:43.5对70.8;p <.001)。与GPT - 4生成的PEMs相比,原始PEMs所需的阅读教育水平更高(平均FKGL:11.6对6.1;p <.001)。对于所有泌尿系统局限性癌症,原始PEMs与GPT - 4生成的PEMs在准确性、完整性和清晰度方面无显著差异。同样,转移性癌症也未观察到差异。GPT生成的PEMs的翻译在77.5%的情况下被评为忠实,在67.5%的情况下被评为清晰。

结论与意义

与原始PEMs相比,GPT - 4生成的PEMs具有更好的可读性水平,同时保持了相似的准确性、完整性和清晰度。利用GenAI的信息提取和语言能力,并结合人工监督,可以显著减少工作量并确保PEMs是最新且准确的。

患者总结

一些为患者编写的癌症相关内容对于前列腺癌、膀胱癌、肾癌或睾丸癌患者来说可能难以阅读或表述不当。本研究使用人工智能从可靠事实中快速生成简短且易于阅读的内容。医生检查了人工智能生成的内容,发现它们与为患者编写的原始文本一样准确、完整且清晰。它们在多种语言中也表现良好。这种人工智能工具可以帮助医疗服务提供者让患者更容易理解自己的癌症以及所能获得的最佳治疗。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/09fc/12136319/83b4f7a247b2/pone.0324175.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/09fc/12136319/372d35dd2c3a/pone.0324175.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/09fc/12136319/0abddbf047f6/pone.0324175.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/09fc/12136319/cf50b469793c/pone.0324175.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/09fc/12136319/83b4f7a247b2/pone.0324175.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/09fc/12136319/372d35dd2c3a/pone.0324175.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/09fc/12136319/0abddbf047f6/pone.0324175.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/09fc/12136319/cf50b469793c/pone.0324175.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/09fc/12136319/83b4f7a247b2/pone.0324175.g004.jpg

相似文献

1
GPT-4 generates accurate and readable patient education materials aligned with current oncological guidelines: A randomized assessment.GPT-4生成符合当前肿瘤学指南的准确且易读的患者教育材料:一项随机评估。
PLoS One. 2025 Jun 4;20(6):e0324175. doi: 10.1371/journal.pone.0324175. eCollection 2025.
2
Readability Assessment of Patient Education Materials on Uro-oncological Diseases Using Automated Measures.使用自动化方法对泌尿肿瘤疾病患者教育材料的可读性评估
Eur Urol Focus. 2024 Dec;10(6):1055-1061. doi: 10.1016/j.euf.2024.06.012. Epub 2024 Jul 23.
3
Accuracy, readability, and understandability of large language models for prostate cancer information to the public.大语言模型向公众提供前列腺癌信息的准确性、可读性和可理解性。
Prostate Cancer Prostatic Dis. 2024 May 14. doi: 10.1038/s41391-024-00826-y.
4
Enhancing the Readability of Online Patient Education Materials Using Large Language Models: Cross-Sectional Study.使用大语言模型提高在线患者教育材料的可读性:横断面研究。
J Med Internet Res. 2025 Jun 4;27:e69955. doi: 10.2196/69955.
5
Empowering patients: how accurate and readable are large language models in renal cancer education.赋能患者:大语言模型在肾癌教育中的准确性和可读性如何。
Front Oncol. 2024 Sep 26;14:1457516. doi: 10.3389/fonc.2024.1457516. eCollection 2024.
6
How readable are Australian multilingual diabetes patient education materials? An evaluation of national English-language source texts.澳大利亚的多语言糖尿病患者教育材料可读性如何?对国家英语源文本的评估。
Public Health Res Pract. 2020 Mar 10;30(1):3012002. doi: 10.17061/phrp3012002.
7
Comparing Artificial Intelligence-Generated and Clinician-Created Personalized Self-Management Guidance for Patients With Knee Osteoarthritis: Blinded Observational Study.比较人工智能生成与临床医生创建的针对膝骨关节炎患者的个性化自我管理指导:盲法观察研究。
J Med Internet Res. 2025 May 7;27:e67830. doi: 10.2196/67830.
8
Evaluation of Generative Language Models in Personalizing Medical Information: Instrument Validation Study.生成式语言模型在个性化医疗信息方面的评估:工具验证研究
JMIR AI. 2024 Aug 13;3:e54371. doi: 10.2196/54371.
9
Leveraging Large Language Models to Enhance Patient Educational Resources in Rhinology.利用大语言模型增强鼻科学中的患者教育资源。
Ann Otol Rhinol Laryngol. 2025 Sep;134(9):683-691. doi: 10.1177/00034894251342969. Epub 2025 May 28.
10
Bridging the Gap Between Urological Research and Patient Understanding: The Role of Large Language Models in Automated Generation of Layperson's Summaries.弥合泌尿科研究与患者理解之间的差距:大型语言模型在生成非专业人士摘要方面的作用。
Urol Pract. 2023 Sep;10(5):436-443. doi: 10.1097/UPJ.0000000000000428. Epub 2023 Jul 5.

引用本文的文献

1
How Accurate Is AI? A Critical Evaluation of Commonly Used Large Language Models in Responding to Patient Concerns About Incidental Kidney Tumors.人工智能的准确性如何?对常用大语言模型回应患者对偶然发现的肾肿瘤担忧的批判性评估。
J Clin Med. 2025 Aug 12;14(16):5697. doi: 10.3390/jcm14165697.
2
[Digital urology : Possible uses for artificial intelligence and digital health applications].[数字泌尿学:人工智能和数字健康应用的可能用途]
Urologie. 2025 Jul 14. doi: 10.1007/s00120-025-02651-0.

本文引用的文献

1
Artificial intelligence improves urologic oncology patient education and counseling.人工智能改善泌尿外科肿瘤患者的教育和咨询。
Can J Urol. 2024 Oct;31(5):12013-12018.
2
The EU Artificial Intelligence Act (2024): Implications for healthcare.欧盟人工智能法案(2024):对医疗保健的影响。
Health Policy. 2024 Nov;149:105152. doi: 10.1016/j.healthpol.2024.105152. Epub 2024 Sep 7.
3
Readability Assessment of Patient Education Materials on Uro-oncological Diseases Using Automated Measures.使用自动化方法对泌尿肿瘤疾病患者教育材料的可读性评估
Eur Urol Focus. 2024 Dec;10(6):1055-1061. doi: 10.1016/j.euf.2024.06.012. Epub 2024 Jul 23.
4
The Use of Artificial Intelligence to Improve Readability of Otolaryngology Patient Education Materials.利用人工智能提高耳鼻喉科患者教育材料的可读性。
Otolaryngol Head Neck Surg. 2024 Aug;171(2):603-608. doi: 10.1002/ohn.816. Epub 2024 May 15.
5
Accuracy, readability, and understandability of large language models for prostate cancer information to the public.大语言模型向公众提供前列腺癌信息的准确性、可读性和可理解性。
Prostate Cancer Prostatic Dis. 2024 May 14. doi: 10.1038/s41391-024-00826-y.
6
Generative artificial intelligence in surgery.手术中的生成式人工智能。
Surgery. 2024 Jun;175(6):1496-1502. doi: 10.1016/j.surg.2024.02.019. Epub 2024 Apr 6.
7
Comment on: Assessing ChatGPT's ability to answer questions pertaining to erectile dysfunction.关于《评估ChatGPT回答与勃起功能障碍相关问题的能力》的评论
Int J Impot Res. 2024 Nov;36(7):796-797. doi: 10.1038/s41443-023-00821-2. Epub 2024 Mar 11.
8
Enfortumab Vedotin and Pembrolizumab in Untreated Advanced Urothelial Cancer.恩福妥单抗Vedotin 联合帕博利珠单抗治疗未经治疗的晚期尿路上皮癌。
N Engl J Med. 2024 Mar 7;390(10):875-888. doi: 10.1056/NEJMoa2312117.
9
Medical Doctors Approaches and Understanding of Health Literacy: A Systematic Literature Review.医生对健康素养的认知与理解:一项系统文献综述
Cureus. 2024 Jan 1;16(1):e51448. doi: 10.7759/cureus.51448. eCollection 2024 Jan.
10
Awareness and Use of ChatGPT and Large Language Models: A Prospective Cross-sectional Global Survey in Urology.泌尿外科中关于 ChatGPT 和大型语言模型的认知和使用:一项前瞻性的全球横断面调查。
Eur Urol. 2024 Feb;85(2):146-153. doi: 10.1016/j.eururo.2023.10.014. Epub 2023 Nov 4.