• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

利用Chat-GPT加强放射科与患者的沟通:评估常见影像相关问题答案的有效性和可读性

Enhancing Patient Communication With Chat-GPT in Radiology: Evaluating the Efficacy and Readability of Answers to Common Imaging-Related Questions.

作者信息

Gordon Emile B, Towbin Alexander J, Wingrove Peter, Shafique Umber, Haas Brian, Kitts Andrea B, Feldman Jill, Furlan Alessandro

机构信息

Department of Radiology, University of Pittsburgh Medical Center, Pittsburgh, Pennsylvania; Clinical Associate, Department of Radiology, Duke University Medical Center, Department of Radiology, Durham, North Carolina.

Professor and Associate Chief, Department of Radiology (Clinical Operations and Informatics), Neil D. Johnson Chair of Radiology Informatics, University of Cincinnati, Cincinnati, Ohio.

出版信息

J Am Coll Radiol. 2024 Feb;21(2):353-359. doi: 10.1016/j.jacr.2023.09.011. Epub 2023 Oct 18.

DOI:10.1016/j.jacr.2023.09.011
PMID:37863153
Abstract

PURPOSE

To assess ChatGPT's accuracy, relevance, and readability in answering patients' common imaging-related questions and examine the effect of a simple prompt.

METHODS

A total of 22 imaging-related questions were developed from categories previously described as important to patients, as follows: safety, the radiology report, the procedure, preparation before imaging, meaning of terms, and medical staff. These questions were posed to ChatGPT with and without a short prompt instructing the model to provide an accurate and easy-to-understand response for the average person. Four board-certified radiologists evaluated the answers for accuracy, consistency, and relevance. Two patient advocates also reviewed responses for their utility for patients. Readability was assessed using the Flesch Kincaid Grade Level. Statistical comparisons were performed using χ and paired t tests.

RESULTS

A total of 264 answers were assessed for both unprompted and prompted questions. Unprompted responses were accurate 83% of the time (218 of 264), which did not significantly change for prompted responses (87% [229 of 264]; P = .2). The consistency of the responses increased from 72% (63 of 88) to 86% (76 of 88) when prompts were given (P = .02). Nearly all responses (99% [261 of 264]) were at least partially relevant for both question types. Fewer unprompted responses were considered fully relevant at 67% (176 of 264), although this increased significantly to 80% when prompts were given (210 of 264; P = .001). The average Flesch Kincaid Grade Level was high at 13.6 [CI, 12.9-14.2], unchanged with the prompt (13.0 [CI, 12.41-13.60], P = .2). None of the responses reached the eighth-grade readability level recommended for patient-facing materials.

DISCUSSION

ChatGPT demonstrates the potential to respond accurately, consistently, and relevantly to patients' imaging-related questions. However, imperfect accuracy and high complexity necessitate oversight before implementation. Prompts reduced response variability and yielded more-targeted information, but they did not improve readability. ChatGPT has the potential to increase accessibility to health information and streamline the production of patient-facing educational materials; however, its current limitations require cautious implementation and further research.

摘要

目的

评估ChatGPT在回答患者常见的影像学相关问题时的准确性、相关性和可读性,并研究一个简单提示的效果。

方法

从先前描述的对患者重要的类别中总共提出了22个影像学相关问题,如下:安全性、放射学报告、检查程序、成像前准备、术语含义和医务人员。这些问题分别在有和没有一个简短提示的情况下向ChatGPT提出,该提示指导模型为普通人提供准确且易于理解的回答。四位获得董事会认证的放射科医生评估答案的准确性、一致性和相关性。两位患者权益倡导者也审查回答对患者的实用性。使用弗莱什-金凯德年级水平评估可读性。使用χ检验和配对t检验进行统计比较。

结果

总共对264个无提示和有提示问题的答案进行了评估。无提示回答的准确率为83%(264个中的218个),有提示回答的准确率没有显著变化(87%[264个中的229个];P = 0.2)。给出提示后,回答的一致性从72%(88个中的63个)提高到86%(88个中的76个)(P = 0.02)。几乎所有回答(99%[264个中的261个])对两种问题类型都至少部分相关。无提示回答中被认为完全相关的较少,为67%(264个中的176个),不过给出提示后这一比例显著提高到80%(264个中的210个;P = 0.001)。平均弗莱什-金凯德年级水平较高,为13.6[置信区间,12.9 - 14.2],有提示时没有变化(13.0[置信区间,12.41 - 13.60],P = 0.2)。没有一个回答达到面向患者材料推荐的八年级可读性水平。

讨论

ChatGPT展示了准确、一致且相关地回答患者影像学相关问题的潜力。然而,准确性不完善和复杂性高需要在实施前进行监督。提示减少了回答的变异性并产生了更有针对性的信息,但没有提高可读性。ChatGPT有潜力增加获取健康信息的机会并简化面向患者的教育材料的制作;然而,其当前的局限性需要谨慎实施和进一步研究。

相似文献

1
Enhancing Patient Communication With Chat-GPT in Radiology: Evaluating the Efficacy and Readability of Answers to Common Imaging-Related Questions.利用Chat-GPT加强放射科与患者的沟通:评估常见影像相关问题答案的有效性和可读性
J Am Coll Radiol. 2024 Feb;21(2):353-359. doi: 10.1016/j.jacr.2023.09.011. Epub 2023 Oct 18.
2
Assessing the Quality and Reliability of ChatGPT's Responses to Radiotherapy-Related Patient Queries: Comparative Study With GPT-3.5 and GPT-4.评估ChatGPT对放疗相关患者问题回答的质量和可靠性:与GPT-3.5和GPT-4的比较研究
JMIR Cancer. 2025 Apr 16;11:e63677. doi: 10.2196/63677.
3
Improving readability in AI-generated medical information on fragility fractures: the role of prompt wording on ChatGPT's responses.提高人工智能生成的关于脆性骨折的医学信息的可读性:提示措辞对ChatGPT回复的作用。
Osteoporos Int. 2025 Mar;36(3):403-410. doi: 10.1007/s00198-024-07358-0. Epub 2025 Jan 8.
4
Assessing the Clinical Appropriateness and Practical Utility of ChatGPT as an Educational Resource for Patients Considering Minimally Invasive Spine Surgery.评估ChatGPT作为考虑微创脊柱手术患者的教育资源的临床适用性和实际效用。
Cureus. 2024 Oct 8;16(10):e71105. doi: 10.7759/cureus.71105. eCollection 2024 Oct.
5
Accuracy and Readability of ChatGPT on Potential Complications of Interventional Radiology Procedures: AI-Powered Patient Interviewing.ChatGPT在介入放射学程序潜在并发症方面的准确性和可读性:人工智能驱动的患者访谈。
Acad Radiol. 2025 Mar;32(3):1547-1553. doi: 10.1016/j.acra.2024.10.028. Epub 2024 Nov 16.
6
Appropriateness and readability of Google Bard and ChatGPT-3.5 generated responses for surgical treatment of glaucoma.谷歌巴德和 ChatGPT-3.5 生成的青光眼手术治疗回复的适宜性和可读性。
Rom J Ophthalmol. 2024 Jul-Sep;68(3):243-248. doi: 10.22336/rjo.2024.45.
7
Information Quality and Readability: ChatGPT's Responses to the Most Common Questions About Spinal Cord Injury.信息质量与可读性:ChatGPT 对脊髓损伤常见问题的回答
World Neurosurg. 2024 Jan;181:e1138-e1144. doi: 10.1016/j.wneu.2023.11.062. Epub 2023 Nov 22.
8
A Multidisciplinary Assessment of ChatGPT's Knowledge of Amyloidosis: Observational Study.对ChatGPT关于淀粉样变性知识的多学科评估:观察性研究。
JMIR Cardio. 2024 Apr 19;8:e53421. doi: 10.2196/53421.
9
Evaluating the accuracy and readability of ChatGPT in providing parental guidance for adenoidectomy, tonsillectomy, and ventilation tube insertion surgery.评估 ChatGPT 在提供腺样体切除术、扁桃体切除术和通气管插入手术的家长指导方面的准确性和可读性。
Int J Pediatr Otorhinolaryngol. 2024 Jun;181:111998. doi: 10.1016/j.ijporl.2024.111998. Epub 2024 May 31.
10
Improving readability of shoulder and elbow surgery online patient education material with Chat GPT (Chat Generative Pretrained Transformer) 4.使用Chat GPT(聊天生成预训练变换器)4提高肩部和肘部手术在线患者教育材料的可读性
J Shoulder Elbow Surg. 2025 Mar 19. doi: 10.1016/j.jse.2025.02.025.

引用本文的文献

1
Artificial Intelligence for Individualized Radiological Dialogue: The Impact of RadioBot on Precision-Driven Medical Practices.用于个性化放射学对话的人工智能:RadioBot对精准驱动医疗实践的影响。
J Pers Med. 2025 Aug 8;15(8):363. doi: 10.3390/jpm15080363.
2
Improving radiology reporting accuracy: use of GPT-4 to reduce errors in reports.提高放射学报告准确性:使用GPT-4减少报告中的错误。
Abdom Radiol (NY). 2025 Jun 27. doi: 10.1007/s00261-025-05079-4.
3
Comparison of ChatGPT and Internet Research for Clinical Research and Decision-Making in Occupational Medicine: Randomized Controlled Trial.
ChatGPT与互联网搜索用于职业医学临床研究和决策的比较:随机对照试验
JMIR Form Res. 2025 May 20;9:e63857. doi: 10.2196/63857.
4
An Australasian survey on the use of ChatGPT and other large language models in medical physics.一项关于在医学物理中使用ChatGPT及其他大语言模型的澳大拉西亚地区调查。
Phys Eng Sci Med. 2025 May 20. doi: 10.1007/s13246-025-01571-9.
5
Artificial intelligence-generated responses to frequently asked questions on coccydynia: Evaluating the accuracy and consistency of GPT-4o's performance.人工智能对尾骨痛常见问题的回答:评估GPT-4o表现的准确性和一致性。
Arch Rheumatol. 2025 Mar 17;40(1):63-71. doi: 10.46497/ArchRheumatol.2025.10966. eCollection 2025 Mar.
6
Enhancing patient-centered information on implant dentistry through prompt engineering: a comparison of four large language models.通过提示工程增强种植牙科以患者为中心的信息:四种大语言模型的比较
Front Oral Health. 2025 Apr 7;6:1566221. doi: 10.3389/froh.2025.1566221. eCollection 2025.
7
Artificial intelligence-large language models (AI-LLMs) for reliable and accurate cardiotocography (CTG) interpretation in obstetric practice.用于产科实践中可靠且准确解读胎心监护(CTG)的人工智能大语言模型(AI-LLMs)。
Comput Struct Biotechnol J. 2025 Mar 18;27:1140-1147. doi: 10.1016/j.csbj.2025.03.026. eCollection 2025.
8
Performance of Large Language Models ChatGPT and Gemini on Workplace Management Questions in Radiology.大型语言模型ChatGPT和Gemini在放射学工作场所管理问题上的表现
Diagnostics (Basel). 2025 Feb 19;15(4):497. doi: 10.3390/diagnostics15040497.
9
Higher education students' perceptions of ChatGPT: A global study of early reactions.高等教育学生对ChatGPT的认知:一项关于早期反应的全球研究。
PLoS One. 2025 Feb 5;20(2):e0315011. doi: 10.1371/journal.pone.0315011. eCollection 2025.
10
Large Language Models for Chatbot Health Advice Studies: A Systematic Review.用于聊天机器人健康建议研究的大语言模型:一项系统综述。
JAMA Netw Open. 2025 Feb 3;8(2):e2457879. doi: 10.1001/jamanetworkopen.2024.57879.