Suppr超能文献

评估人工智能对泌尿科患者收件箱信息的回复。

Assessing Artificial Intelligence-Generated Responses to Urology Patient In-Basket Messages.

机构信息

Department of Urology, Stanford University School of Medicine, Palo Alto, California.

Idaho Urologic Institute, Meridian, Idaho.

出版信息

Urol Pract. 2024 Sep;11(5):793-798. doi: 10.1097/UPJ.0000000000000637. Epub 2024 Jun 24.

Abstract

INTRODUCTION

Electronic patient messaging utilization has increased in recent years and has been associated with physician burnout. ChatGPT is a language model that has shown the ability to generate near-human level text responses. This study evaluated the quality of ChatGPT responses to real-world urology patient messages.

METHODS

One hundred electronic patient messages were collected from a practicing urologist's inbox and categorized based on the question content. Individual responses were generated by entering each message into ChatGPT. The questions and responses were independently evaluated by 5 urologists and graded on a 5-point Likert scale. Questions were graded based on difficulty, and responses were graded based on accuracy, completeness, harmfulness, helpfulness, and intelligibleness. Whether or not the response could be sent to a patient was also assessed.

RESULTS

Overall, 47% of responses were deemed acceptable to send to patients. ChatGPT performed better on easy questions with 56% of responses to easy questions being acceptable to send as compared to 34% of difficult questions ( = .03). Responses to easy questions were more accurate, complete, helpful, and intelligible than responses to difficult questions. There was no difference in response quality based on question content.

CONCLUSIONS

ChatGPT generated acceptable responses to nearly 50% of patient messages with better performance for easy questions compared to difficult questions. Use of ChatGPT to help respond to patient messages can help to decrease the time burden for the care team and improve wellness. Artificial intelligence performance will likely continue to improve with advances in generative artificial intelligence technology.

摘要

简介

近年来,电子患者信息传递的使用有所增加,并且与医生倦怠有关。ChatGPT 是一种语言模型,它已经显示出生成接近人类水平文本回复的能力。本研究评估了 ChatGPT 对真实世界泌尿科患者信息的回复质量。

方法

从一名执业泌尿科医生的收件箱中收集了 100 封电子患者信息,并根据问题内容进行了分类。通过将每条信息输入 ChatGPT 来生成单独的回复。5 位泌尿科医生对问题和回复进行了独立评估,并按照 5 分制的李克特量表进行了评分。根据问题的难度对问题进行评分,根据准确性、完整性、危害性、有用性和可理解性对回复进行评分。还评估了回复是否可以发送给患者。

结果

总体而言,47%的回复被认为可以发送给患者。ChatGPT 在简单问题上的表现更好,56%的简单问题回复可以发送,而 34%的困难问题回复可以发送( =.03)。与困难问题的回复相比,简单问题的回复更准确、完整、有用和易懂。根据问题内容,回复质量没有差异。

结论

ChatGPT 对近 50%的患者信息生成了可接受的回复,与困难问题相比,简单问题的性能更好。使用 ChatGPT 来帮助回复患者信息可以帮助减轻护理团队的时间负担并提高健康水平。随着生成式人工智能技术的进步,人工智能的性能可能会继续提高。

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验