• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

评估人工智能聊天机器人在传递 IR 患者教育材料方面的效果:与社会网站内容的比较。

Evaluation of an Artificial Intelligence Chatbot for Delivery of IR Patient Education Material: A Comparison with Societal Website Content.

机构信息

Division of Vascular and Interventional Radiology, Beth Israel Deaconess Medical Center, Harvard Medical School, Boston, Massachusetts.

Division of Vascular and Interventional Radiology, Beth Israel Deaconess Medical Center, Harvard Medical School, Boston, Massachusetts.

出版信息

J Vasc Interv Radiol. 2023 Oct;34(10):1760-1768.e32. doi: 10.1016/j.jvir.2023.05.037. Epub 2023 Jun 16.

DOI:10.1016/j.jvir.2023.05.037
PMID:37330210
Abstract

PURPOSE

To assess the accuracy, completeness, and readability of patient educational material produced by a machine learning model and compare the output to that provided by a societal website.

MATERIALS AND METHODS

Content from the Society of Interventional Radiology Patient Center website was retrieved, categorized, and organized into discrete questions. These questions were entered into the ChatGPT platform, and the output was analyzed for word and sentence counts, readability using multiple validated scales, factual correctness, and suitability for patient education using the Patient Education Materials Assessment Tool for Printable Materials (PEMAT-P) instrument.

RESULTS

A total of 21,154 words were analyzed, including 7,917 words from the website and 13,377 words representing the total output of the ChatGPT platform across 22 text passages. Compared to the societal website, output from the ChatGPT platform was longer and more difficult to read on 4 of 5 readability scales. The ChatGPT output was incorrect for 12 (11.5%) of 104 questions. When reviewed using the PEMAT-P tool, the ChatGPT content scored lower than the website material. Content from both the website and ChatGPT were significantly above the recommended fifth or sixth grade level for patient education, with a mean Flesch-Kincaid grade level of 11.1 (±1.3) for the website and 11.9 (±1.6) for the ChatGPT content.

CONCLUSIONS

The ChatGPT platform may produce incomplete or inaccurate patient educational content, and providers should be familiar with the limitations of the system in its current form. Opportunities may exist to fine-tune existing large language models, which could be optimized for the delivery of patient educational content.

摘要

目的

评估机器学习模型生成的患者教育材料的准确性、完整性和可读性,并将其输出与社会网站提供的内容进行比较。

材料与方法

从介入放射学学会患者中心网站检索内容,进行分类,并组织成离散的问题。将这些问题输入 ChatGPT 平台,分析其词汇和句子数量、使用多个经过验证的量表进行的可读性、事实正确性以及使用印刷材料患者教育材料评估工具(PEMAT-P)仪器进行患者教育的适宜性。

结果

共分析了 21154 个单词,包括网站的 7917 个单词和 ChatGPT 平台的 13377 个单词,涵盖了 22 个文本段落的总输出。与社会网站相比,ChatGPT 平台的输出在 5 个可读性量表中的 4 个上更长且更难阅读。ChatGPT 输出在 104 个问题中的 12 个(11.5%)是错误的。使用 PEMAT-P 工具进行审查时,ChatGPT 内容的评分低于网站材料。网站和 ChatGPT 的内容都明显高于推荐的五年级或六年级的患者教育水平,网站的 Flesch-Kincaid 平均等级为 11.1(±1.3),ChatGPT 内容的 Flesch-Kincaid 平均等级为 11.9(±1.6)。

结论

ChatGPT 平台可能会生成不完整或不准确的患者教育内容,并且提供者应该熟悉其当前形式的系统限制。可能存在机会微调现有的大型语言模型,使其可以优化用于提供患者教育内容。

相似文献

1
Evaluation of an Artificial Intelligence Chatbot for Delivery of IR Patient Education Material: A Comparison with Societal Website Content.评估人工智能聊天机器人在传递 IR 患者教育材料方面的效果:与社会网站内容的比较。
J Vasc Interv Radiol. 2023 Oct;34(10):1760-1768.e32. doi: 10.1016/j.jvir.2023.05.037. Epub 2023 Jun 16.
2
A quality and readability comparison of artificial intelligence and popular health website education materials for common hand surgery procedures.常见手部手术程序的人工智能与热门健康网站教育资料的质量和可读性比较。
Hand Surg Rehabil. 2024 Jun;43(3):101723. doi: 10.1016/j.hansur.2024.101723. Epub 2024 May 21.
3
Dr. Google vs. Dr. ChatGPT: Exploring the Use of Artificial Intelligence in Ophthalmology by Comparing the Accuracy, Safety, and Readability of Responses to Frequently Asked Patient Questions Regarding Cataracts and Cataract Surgery.谷歌医生与ChatGPT医生:通过比较关于白内障及白内障手术的常见患者问题的回答的准确性、安全性和可读性,探索人工智能在眼科领域的应用。
Semin Ophthalmol. 2024 Aug;39(6):472-479. doi: 10.1080/08820538.2024.2326058. Epub 2024 Mar 22.
4
Evaluating ChatGPT platform in delivering heart failure educational material: A comparison with the leading national cardiology institutes.评估 ChatGPT 平台在传递心力衰竭教育材料方面的效果:与领先的国家心脏病学研究所的比较。
Curr Probl Cardiol. 2024 Nov;49(11):102797. doi: 10.1016/j.cpcardiol.2024.102797. Epub 2024 Aug 17.
5
Comparison of ChatGPT and Traditional Patient Education Materials for Men's Health.比较 ChatGPT 和传统男性健康患者教育材料。
Urol Pract. 2024 Jan;11(1):87-94. doi: 10.1097/UPJ.0000000000000490. Epub 2023 Nov 1.
6
Artificial intelligence chatbots as sources of patient education material for obstructive sleep apnoea: ChatGPT versus Google Bard.人工智能聊天机器人作为阻塞性睡眠呼吸暂停患者教育材料的来源:ChatGPT 与 Google Bard 对比。
Eur Arch Otorhinolaryngol. 2024 Feb;281(2):985-993. doi: 10.1007/s00405-023-08319-9. Epub 2023 Nov 2.
7
ChatGPT vs. web search for patient questions: what does ChatGPT do better?ChatGPT 与网页搜索在解答患者问题上的对比:ChatGPT 有哪些优势?
Eur Arch Otorhinolaryngol. 2024 Jun;281(6):3219-3225. doi: 10.1007/s00405-024-08524-0. Epub 2024 Feb 28.
8
The Performance of Chatbots and the AAPOS Website as a Tool for Amblyopia Education.聊天机器人和 AAPOS 网站在弱视教育中的应用效果。
J Pediatr Ophthalmol Strabismus. 2024 Sep-Oct;61(5):325-331. doi: 10.3928/01913913-20240409-01. Epub 2024 May 30.
9
Assessing the Application of Large Language Models in Generating Dermatologic Patient Education Materials According to Reading Level: Qualitative Study.评估大语言模型在根据阅读水平生成皮肤科患者教育材料方面的应用:定性研究。
JMIR Dermatol. 2024 May 16;7:e55898. doi: 10.2196/55898.
10
Talking technology: exploring chatbots as a tool for cataract patient education.技术漫谈:探索聊天机器人作为白内障患者教育工具的作用
Clin Exp Optom. 2025 Jan;108(1):56-64. doi: 10.1080/08164622.2023.2298812. Epub 2024 Jan 9.

引用本文的文献

1
Comparison of the readability of ChatGPT and Bard in medical communication: a meta-analysis.ChatGPT与Bard在医学交流中的可读性比较:一项荟萃分析。
BMC Med Inform Decis Mak. 2025 Sep 1;25(1):325. doi: 10.1186/s12911-025-03035-2.
2
Evaluating the readability, quality, and reliability of responses generated by ChatGPT, Gemini, and Perplexity on the most commonly asked questions about Ankylosing spondylitis.评估ChatGPT、Gemini和Perplexity针对强直性脊柱炎最常见问题生成的回答的可读性、质量和可靠性。
PLoS One. 2025 Jun 18;20(6):e0326351. doi: 10.1371/journal.pone.0326351. eCollection 2025.
3
ChatGPT-4 Effectively Responds to Common Patient Questions on Total Ankle Arthroplasty: A Surgeon-Based Assessment of AI in Patient Education.
ChatGPT-4有效回应患者关于全踝关节置换术的常见问题:基于外科医生的人工智能在患者教育中的评估
Foot Ankle Orthop. 2025 Mar 27;10(1):24730114251322784. doi: 10.1177/24730114251322784. eCollection 2025 Jan.
4
Evaluating the Use of Generative Artificial Intelligence to Support Genetic Counseling for Rare Diseases.评估生成式人工智能在支持罕见病遗传咨询中的应用。
Diagnostics (Basel). 2025 Mar 10;15(6):672. doi: 10.3390/diagnostics15060672.
5
Large language model doctor: assessing the ability of ChatGPT-4 to deliver interventional radiology procedural information to patients during the consent process.大语言模型医生:评估ChatGPT-4在知情同意过程中向患者提供介入放射学程序信息的能力。
CVIR Endovasc. 2024 Nov 29;7(1):83. doi: 10.1186/s42155-024-00477-z.
6
Large language models in patient education: a scoping review of applications in medicine.用于患者教育的大语言模型:医学应用的范围综述
Front Med (Lausanne). 2024 Oct 29;11:1477898. doi: 10.3389/fmed.2024.1477898. eCollection 2024.
7
Development and Validation of an Artificial Intelligence-Assisted Patient Education Material for Ostomy Patients: A Methodological Study.造口患者人工智能辅助患者教育材料的开发与验证:一项方法学研究。
J Adv Nurs. 2025 Jul;81(7):3859-3867. doi: 10.1111/jan.16542. Epub 2024 Oct 18.
8
Comparative analysis of GPT-4-based ChatGPT's diagnostic performance with radiologists using real-world radiology reports of brain tumors.基于GPT-4的ChatGPT与放射科医生在使用脑肿瘤真实世界放射学报告方面的诊断性能比较分析。
Eur Radiol. 2025 Apr;35(4):1938-1947. doi: 10.1007/s00330-024-11032-8. Epub 2024 Aug 28.
9
Generative Artificial Intelligence: Enhancing Patient Education in Cardiovascular Imaging.生成式人工智能:加强心血管成像中的患者教育
BJR Open. 2024 Jul 17;6(1):tzae018. doi: 10.1093/bjro/tzae018. eCollection 2024 Jan.
10
Using artificial intelligence to generate medical literature for urology patients: a comparison of three different large language models.利用人工智能为泌尿科患者生成医学文献:三种不同的大型语言模型比较。
World J Urol. 2024 Jul 29;42(1):455. doi: 10.1007/s00345-024-05146-3.