• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

患者群体对 GPT 生成的与标准急诊部门出院医嘱的看法:随机盲法调查评估。

Patient-Representing Population's Perceptions of GPT-Generated Versus Standard Emergency Department Discharge Instructions: Randomized Blind Survey Assessment.

机构信息

Department of Emergency Medicine, Yale School of Medicine, New Haven, CT, United States.

Department for Biomedical Informatics and Data Science, Yale School of Medicine, New Haven, CT, United States.

出版信息

J Med Internet Res. 2024 Aug 2;26:e60336. doi: 10.2196/60336.

DOI:10.2196/60336
PMID:39094112
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11329854/
Abstract

BACKGROUND

Discharge instructions are a key form of documentation and patient communication in the time of transition from the emergency department (ED) to home. Discharge instructions are time-consuming and often underprioritized, especially in the ED, leading to discharge delays and possibly impersonal patient instructions. Generative artificial intelligence and large language models (LLMs) offer promising methods of creating high-quality and personalized discharge instructions; however, there exists a gap in understanding patient perspectives of LLM-generated discharge instructions.

OBJECTIVE

We aimed to assess the use of LLMs such as ChatGPT in synthesizing accurate and patient-accessible discharge instructions in the ED.

METHODS

We synthesized 5 unique, fictional ED encounters to emulate real ED encounters that included a diverse set of clinician history, physical notes, and nursing notes. These were passed to GPT-4 in Azure OpenAI Service (Microsoft) to generate LLM-generated discharge instructions. Standard discharge instructions were also generated for each of the 5 unique ED encounters. All GPT-generated and standard discharge instructions were then formatted into standardized after-visit summary documents. These after-visit summaries containing either GPT-generated or standard discharge instructions were randomly and blindly administered to Amazon MTurk respondents representing patient populations through Amazon MTurk Survey Distribution. Discharge instructions were assessed based on metrics of interpretability of significance, understandability, and satisfaction.

RESULTS

Our findings revealed that survey respondents' perspectives regarding GPT-generated and standard discharge instructions were significantly (P=.01) more favorable toward GPT-generated return precautions, and all other sections were considered noninferior to standard discharge instructions. Of the 156 survey respondents, GPT-generated discharge instructions were assigned favorable ratings, "agree" and "strongly agree," more frequently along the metric of interpretability of significance in discharge instruction subsections regarding diagnosis, procedures, treatment, post-ED medications or any changes to medications, and return precautions. Survey respondents found GPT-generated instructions to be more understandable when rating procedures, treatment, post-ED medications or medication changes, post-ED follow-up, and return precautions. Satisfaction with GPT-generated discharge instruction subsections was the most favorable in procedures, treatment, post-ED medications or medication changes, and return precautions. Wilcoxon rank-sum test of Likert responses revealed significant differences (P=.01) in the interpretability of significant return precautions in GPT-generated discharge instructions compared to standard discharge instructions but not for other evaluation metrics and discharge instruction subsections.

CONCLUSIONS

This study demonstrates the potential for LLMs such as ChatGPT to act as a method of augmenting current documentation workflows in the ED to reduce the documentation burden of physicians. The ability of LLMs to provide tailored instructions for patients by improving readability and making instructions more applicable to patients could improve upon the methods of communication that currently exist.

摘要

背景

在从急诊科(ED)过渡到家庭的过程中,出院说明是一种关键的文档和患者沟通形式。出院说明既耗时又常常被优先考虑,尤其是在急诊科,这导致了出院延迟和可能不人性化的患者说明。生成式人工智能和大型语言模型(LLM)为创建高质量和个性化的出院说明提供了有前途的方法;然而,在理解患者对 LLM 生成的出院说明的看法方面仍存在差距。

目的

我们旨在评估在急诊科使用 ChatGPT 等 LLM 来综合准确且易于患者理解的出院说明。

方法

我们综合了 5 个独特的虚构 ED 就诊,以模拟包括一系列临床医生病史、体检记录和护理记录的真实 ED 就诊。这些信息被传递给 Azure OpenAI Service(微软)中的 GPT-4 以生成 LLM 生成的出院说明。每个独特的 ED 就诊也生成了标准的出院说明。所有 GPT 生成的和标准的出院说明随后都被格式化为标准化的就诊后总结文档。这些就诊后总结文档包含 GPT 生成或标准的出院说明,通过亚马逊 MTurk 调查分配随机和盲目地分发给代表患者群体的亚马逊 MTurk 受访者。根据可解释性、理解性和满意度的指标来评估出院说明。

结果

我们的研究结果表明,调查受访者对 GPT 生成和标准出院说明的看法在出院后预防措施方面有显著差异(P=.01),并且所有其他部分都被认为与标准出院说明无差异。在 156 名调查受访者中,GPT 生成的出院说明在诊断、程序、治疗、ED 后药物或任何药物变化以及出院后预防措施等出院说明子部分的意义可解释性方面,更频繁地被评为“同意”和“强烈同意”。调查受访者发现,在程序、治疗、ED 后药物或药物变化、ED 后随访以及出院后预防措施方面,GPT 生成的说明更容易理解。GPT 生成的出院说明子部分的满意度在程序、治疗、ED 后药物或药物变化以及出院后预防措施方面最高。对 GPT 生成的出院说明子部分的李克特反应进行 Wilcoxon 秩和检验,结果显示,在 GPT 生成的出院说明中,与标准出院说明相比,在意义重大的出院后预防措施的可解释性方面存在显著差异(P=.01),但其他评估指标和出院说明子部分则没有。

结论

本研究表明,像 ChatGPT 这样的 LLM 可以作为一种在急诊科中增强当前文档工作流程的方法,以减轻医生的文档负担。LLM 通过提高可读性和使说明更适用于患者,为患者提供量身定制的说明的能力,可以改进当前存在的沟通方法。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/768c/11329854/dd4a0c01a956/jmir_v26i1e60336_fig5.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/768c/11329854/a59ddbb8fa2d/jmir_v26i1e60336_fig1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/768c/11329854/3531fdc678ac/jmir_v26i1e60336_fig2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/768c/11329854/80775eccc83b/jmir_v26i1e60336_fig3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/768c/11329854/3044ed172330/jmir_v26i1e60336_fig4.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/768c/11329854/dd4a0c01a956/jmir_v26i1e60336_fig5.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/768c/11329854/a59ddbb8fa2d/jmir_v26i1e60336_fig1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/768c/11329854/3531fdc678ac/jmir_v26i1e60336_fig2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/768c/11329854/80775eccc83b/jmir_v26i1e60336_fig3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/768c/11329854/3044ed172330/jmir_v26i1e60336_fig4.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/768c/11329854/dd4a0c01a956/jmir_v26i1e60336_fig5.jpg

相似文献

1
Patient-Representing Population's Perceptions of GPT-Generated Versus Standard Emergency Department Discharge Instructions: Randomized Blind Survey Assessment.患者群体对 GPT 生成的与标准急诊部门出院医嘱的看法:随机盲法调查评估。
J Med Internet Res. 2024 Aug 2;26:e60336. doi: 10.2196/60336.
2
Evaluating Large Language Models for Drafting Emergency Department Discharge Summaries.评估用于起草急诊科出院小结的大语言模型。
medRxiv. 2024 Apr 4:2024.04.03.24305088. doi: 10.1101/2024.04.03.24305088.
3
AI in Home Care-Evaluation of Large Language Models for Future Training of Informal Caregivers: Observational Comparative Case Study.家庭护理中的人工智能——对用于未来非正式护理人员培训的大语言模型的评估:观察性比较案例研究
J Med Internet Res. 2025 Apr 28;27:e70703. doi: 10.2196/70703.
4
ChatGPT With GPT-4 Outperforms Emergency Department Physicians in Diagnostic Accuracy: Retrospective Analysis.ChatGPT 联合 GPT-4 在诊断准确率上优于急诊科医生:回顾性分析。
J Med Internet Res. 2024 Jul 8;26:e56110. doi: 10.2196/56110.
5
Triage Performance Across Large Language Models, ChatGPT, and Untrained Doctors in Emergency Medicine: Comparative Study.分诊表现比较:大型语言模型、ChatGPT 和未经训练的急诊医生:一项对比研究。
J Med Internet Res. 2024 Jun 14;26:e53297. doi: 10.2196/53297.
6
Generative Artificial Intelligence to Transform Inpatient Discharge Summaries to Patient-Friendly Language and Format.生成式人工智能将住院病历摘要转换为患者友好型语言和格式。
JAMA Netw Open. 2024 Mar 4;7(3):e240357. doi: 10.1001/jamanetworkopen.2024.0357.
7
Virtual Patients Using Large Language Models: Scalable, Contextualized Simulation of Clinician-Patient Dialogue With Feedback.使用大语言模型的虚拟患者:具有反馈功能的临床医生-患者对话的可扩展、情境化模拟
J Med Internet Res. 2025 Apr 4;27:e68486. doi: 10.2196/68486.
8
Assessing the Application of Large Language Models in Generating Dermatologic Patient Education Materials According to Reading Level: Qualitative Study.评估大语言模型在根据阅读水平生成皮肤科患者教育材料方面的应用:定性研究。
JMIR Dermatol. 2024 May 16;7:e55898. doi: 10.2196/55898.
9
Quality of Answers of Generative Large Language Models Versus Peer Users for Interpreting Laboratory Test Results for Lay Patients: Evaluation Study.生成式大语言模型与同行用户对解释非专业患者实验室检测结果的答案质量比较:评估研究。
J Med Internet Res. 2024 Apr 17;26:e56655. doi: 10.2196/56655.
10
Comparing Artificial Intelligence-Generated and Clinician-Created Personalized Self-Management Guidance for Patients With Knee Osteoarthritis: Blinded Observational Study.比较人工智能生成与临床医生创建的针对膝骨关节炎患者的个性化自我管理指导:盲法观察研究。
J Med Internet Res. 2025 May 7;27:e67830. doi: 10.2196/67830.

引用本文的文献

1
Evaluation of the accuracy and safety of machine translation of patient-specific discharge instructions: a comparative analysis.患者特定出院指导机器翻译的准确性和安全性评估:一项对比分析。
BMJ Qual Saf. 2025 Jul 9. doi: 10.1136/bmjqs-2024-018384.
2
ChatGPT Applications in Nursing: Current Status and Future Perspectives.ChatGPT在护理中的应用:现状与未来展望。
Nurs Open. 2025 Jun;12(6):e70253. doi: 10.1002/nop2.70253.
3
A Practical Guide to the Utilization of ChatGPT in the Emergency Department: A Systematic Review of Current Applications, Future Directions, and Limitations.

本文引用的文献

1
ChatGPT-4 generates orthopedic discharge documents faster than humans maintaining comparable quality: a pilot study of 6 cases.ChatGPT-4 生成骨科出院文件的速度快于人类,且保持相同质量:6 例初步研究。
Acta Orthop. 2024 Mar 21;95:152-156. doi: 10.2340/17453674.2024.40182.
2
Harnessing the Power of Generative AI for Clinical Summaries: Perspectives From Emergency Physicians.利用生成式人工智能为临床总结提供助力:来自急诊医师的观点。
Ann Emerg Med. 2024 Aug;84(2):128-138. doi: 10.1016/j.annemergmed.2024.01.039. Epub 2024 Mar 12.
3
Generative Artificial Intelligence to Transform Inpatient Discharge Summaries to Patient-Friendly Language and Format.
急诊科使用ChatGPT实用指南:当前应用、未来方向及局限性的系统评价
Cureus. 2025 Apr 6;17(4):e81802. doi: 10.7759/cureus.81802. eCollection 2025 Apr.
4
Parental, Patient, and Provider Perspectives on the Use of Artificial Intelligence in Pediatric Care.家长、患者及医疗服务提供者对人工智能在儿科护理中应用的看法。
Acad Pediatr. 2025 Aug;25(6):102819. doi: 10.1016/j.acap.2025.102819. Epub 2025 Mar 22.
5
Artificial intelligence, medications, pharmacogenomics, and ethics.人工智能、药物、药物基因组学与伦理学。
Pharmacogenomics. 2024;25(14-15):611-622. doi: 10.1080/14622416.2024.2428587. Epub 2024 Nov 15.
生成式人工智能将住院病历摘要转换为患者友好型语言和格式。
JAMA Netw Open. 2024 Mar 4;7(3):e240357. doi: 10.1001/jamanetworkopen.2024.0357.
4
Large language models: a primer and gastroenterology applications.大语言模型:入门介绍及胃肠病学应用
Therap Adv Gastroenterol. 2024 Feb 22;17:17562848241227031. doi: 10.1177/17562848241227031. eCollection 2024.
5
Learning to Make Rare and Complex Diagnoses With Generative AI Assistance: Qualitative Study of Popular Large Language Models.利用生成式人工智能辅助学习罕见且复杂的诊断:对流行的大型语言模型的定性研究。
JMIR Med Educ. 2024 Feb 13;10:e51391. doi: 10.2196/51391.
6
Accuracy of ChatGPT, Google Bard, and Microsoft Bing for Simplifying Radiology Reports.ChatGPT、谷歌巴德和微软必应简化放射学报告的准确性。
Radiology. 2023 Nov;309(2):e232561. doi: 10.1148/radiol.232561.
7
A Comprehensive Survey of ChatGPT: Advancements, Applications, Prospects, and Challenges.ChatGPT综合调查:进展、应用、前景与挑战
Meta Radiol. 2023 Sep;1(2). doi: 10.1016/j.metrad.2023.100022. Epub 2023 Oct 7.
8
Comparing ChatGPT and GPT-4 performance in USMLE soft skill assessments.比较 ChatGPT 和 GPT-4 在 USMLE 软技能评估中的表现。
Sci Rep. 2023 Oct 1;13(1):16492. doi: 10.1038/s41598-023-43436-9.
9
A method to automate the discharge summary hospital course for neurology patients.一种自动化神经内科患者出院小结住院流程的方法。
J Am Med Inform Assoc. 2023 Nov 17;30(12):1995-2003. doi: 10.1093/jamia/ocad177.
10
Comparison of History of Present Illness Summaries Generated by a Chatbot and Senior Internal Medicine Residents.聊天机器人与内科住院医师生成的现病史摘要比较
JAMA Intern Med. 2023 Sep 1;183(9):1026-1027. doi: 10.1001/jamainternmed.2023.2561.