Suppr超能文献

评估人工智能驱动的患者教育:放射学案例研究。

Assessing AI-Powered Patient Education: A Case Study in Radiology.

机构信息

University of Wisconsin School of Medicine and Public Health, 750 Highland Ave, Madison, WI 53705.

University of Maryland School of Medicine, Baltimore, Maryland.

出版信息

Acad Radiol. 2024 Jan;31(1):338-342. doi: 10.1016/j.acra.2023.08.020. Epub 2023 Sep 14.

Abstract

RATIONALE AND OBJECTIVES

With recent advancements in the power and accessibility of artificial intelligence (AI) Large Language Models (LLMs) patients might increasingly turn to these platforms to answer questions regarding radiologic examinations and procedures, despite valid concerns about the accuracy of information provided. This study aimed to assess the accuracy and completeness of information provided by the Bing Chatbot-a LLM powered by ChatGPT-on patient education for common radiologic exams.

MATERIALS AND METHODS

We selected three common radiologic examinations and procedures: computed tomography (CT) abdomen, magnetic resonance imaging (MRI) spine, and bone biopsy. For each, ten questions were tested on the chatbot in two trials using three different chatbot settings. Two reviewers independently assessed the chatbot's responses for accuracy and completeness compared to an accepted online resource, radiologyinfo.org.

RESULTS

Of the 360 reviews performed, 336 (93%) were rated "entirely correct" and 24 (7%) were "mostly correct," indicating a high level of reliability. Completeness ratings showed that 65% were "complete" and 35% were "mostly complete." The "More Creative" chatbot setting produced a higher proportion of responses rated "entirely correct" but there were otherwise no significant difference in ratings based on chatbot settings or exam types. The readability level was rated eighth-grade level.

CONCLUSION

The Bing Chatbot provided accurate responses answering all or most aspects of the question asked of it, with responses tending to err on the side of caution for nuanced questions. Importantly, no responses were inaccurate or had potential to cause harm or confusion for the user. Thus, LLM chatbots demonstrate potential to enhance patient education in radiology and could be integrated into patient portals for various purposes, including exam preparation and results interpretation.

摘要

背景与目的

随着人工智能(AI)大型语言模型(LLM)的功能和普及度的提高,患者可能会越来越多地转向这些平台来获取有关放射学检查和操作的问题答案,尽管人们对提供信息的准确性存在合理的担忧。本研究旨在评估由 ChatGPT 驱动的 Bing Chatbot 为常见放射学检查提供的患者教育信息的准确性和完整性。

材料与方法

我们选择了三种常见的放射学检查和操作:腹部 CT、脊柱 MRI 和骨活检。对于每种检查和操作,我们在两次试验中使用三种不同的聊天机器人设置在聊天机器人上测试了十个问题。两名审查员独立评估了聊天机器人的回答与接受的在线资源 radiologyinfo.org 的准确性和完整性。

结果

在进行的 360 次评估中,336 次(93%)被评为“完全正确”,24 次(7%)被评为“大部分正确”,表明可靠性较高。完整性评级显示,65%是“完整”的,35%是“大部分完整”的。“更具创意”的聊天机器人设置产生了更高比例的被评为“完全正确”的回答,但在基于聊天机器人设置或检查类型的评级方面没有其他显著差异。可读性水平被评为八年级水平。

结论

Bing Chatbot 提供了准确的回答,回答了提出的所有或大部分问题,回答倾向于对细微问题持谨慎态度。重要的是,没有回答不准确或有可能对用户造成伤害或混淆。因此,大型语言模型聊天机器人有可能增强放射学患者教育,并可整合到患者门户中用于各种目的,包括检查准备和结果解释。

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验