• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

评估人工智能在核心脏病学方面的能力:大型语言模型参加资格考试准备。

Evaluating AI Proficiency in Nuclear Cardiology: Large Language Models take on the Board Preparation Exam.

作者信息

Builoff Valerie, Shanbhag Aakash, Miller Robert Jh, Dey Damini, Liang Joanna X, Flood Kathleen, Bourque Jamieson M, Chareonthaitawee Panithaya, Phillips Lawrence M, Slomka Piotr J

机构信息

Departments of Medicine (Division of Artificial Intelligence in Medicine), Imaging and Biomedical Sciences Cedars-Sinai Medical Center, Los Angeles, CA, USA.

Signal and Image Processing Institute, Ming Hsieh Department of Electrical and Computer Engineering, University of Southern California, Los Angeles, CA, USA.

出版信息

medRxiv. 2024 Jul 16:2024.07.16.24310297. doi: 10.1101/2024.07.16.24310297.

DOI:10.1101/2024.07.16.24310297
PMID:39072028
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11275690/
Abstract

BACKGROUND

Previous studies evaluated the ability of large language models (LLMs) in medical disciplines; however, few have focused on image analysis, and none specifically on cardiovascular imaging or nuclear cardiology.

OBJECTIVES

This study assesses four LLMs - GPT-4, GPT-4 Turbo, GPT-4omni (GPT-4o) (Open AI), and Gemini (Google Inc.) - in responding to questions from the 2023 American Society of Nuclear Cardiology Board Preparation Exam, reflecting the scope of the Certification Board of Nuclear Cardiology (CBNC) examination.

METHODS

We used 168 questions: 141 text-only and 27 image-based, categorized into four sections mirroring the CBNC exam. Each LLM was presented with the same standardized prompt and applied to each section 30 times to account for stochasticity. Performance over six weeks was assessed for all models except GPT-4o. McNemar's test compared correct response proportions.

RESULTS

GPT-4, Gemini, GPT4-Turbo, and GPT-4o correctly answered median percentiles of 56.8% (95% confidence interval 55.4% - 58.0%), 40.5% (39.9% - 42.9%), 60.7% (59.9% - 61.3%) and 63.1% (62.5 - 64.3%) of questions, respectively. GPT4o significantly outperformed other models (p=0.007 vs. GPT-4Turbo, p<0.001 vs. GPT-4 and Gemini). GPT-4o excelled on text-only questions compared to GPT-4, Gemini, and GPT-4 Turbo (p<0.001, p<0.001, and p=0.001), while Gemini performed worse on image-based questions (p<0.001 for all).

CONCLUSION

GPT-4o demonstrated superior performance among the four LLMs, achieving scores likely within or just outside the range required to pass a test akin to the CBNC examination. Although improvements in medical image interpretation are needed, GPT-4o shows potential to support physicians in answering text-based clinical questions.

摘要

背景

以往的研究评估了大语言模型(LLMs)在医学领域的能力;然而,很少有研究关注图像分析,且没有专门针对心血管成像或核心脏病学的研究。

目的

本研究评估了四种大语言模型——GPT-4、GPT-4 Turbo、GPT-4omni(GPT-4o)(OpenAI)和Gemini(谷歌公司)——对2023年美国核心脏病学会委员会预备考试问题的回答情况,该考试反映了核心脏病学认证委员会(CBNC)考试的范围。

方法

我们使用了168个问题:141个纯文本问题和27个基于图像的问题,分为四个部分,与CBNC考试相对应。每个大语言模型都收到相同的标准化提示,并应用于每个部分30次,以考虑随机性。除GPT-4o外,对所有模型在六周内的表现进行了评估。使用McNemar检验比较正确回答比例。

结果

GPT-4、Gemini、GPT4-Turbo和GPT-4o正确回答问题的中位数百分比分别为56.8%(95%置信区间55.4% - 58.0%)、40.5%(39.9% - 42.9%)、60.7%(59.9% - 61.3%)和63.1%(62.5 - 64.3%)。GPT4o显著优于其他模型(与GPT-4Turbo相比,p = 0.007;与GPT-4和Gemini相比,p < 0.001)。与GPT-4、Gemini和GPT-4 Turbo相比,GPT-4o在纯文本问题上表现出色(p < 0.001、p < 0.001和p = 0.001),而Gemini在基于图像的问题上表现较差(所有比较p < 0.001)。

结论

GPT-4o在这四种大语言模型中表现出卓越的性能,其得分可能在类似于CBNC考试的及格范围内或略高于该范围。尽管医学图像解读仍需改进,但GPT-4o显示出支持医生回答基于文本的临床问题的潜力。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9ab7/11275690/26e2cdc8ce7f/nihpp-2024.07.16.24310297v1-f0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9ab7/11275690/acf5c69148bb/nihpp-2024.07.16.24310297v1-f0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9ab7/11275690/5615512b0b4c/nihpp-2024.07.16.24310297v1-f0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9ab7/11275690/3aad29aa1dbe/nihpp-2024.07.16.24310297v1-f0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9ab7/11275690/f9e54ecb0684/nihpp-2024.07.16.24310297v1-f0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9ab7/11275690/26e2cdc8ce7f/nihpp-2024.07.16.24310297v1-f0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9ab7/11275690/acf5c69148bb/nihpp-2024.07.16.24310297v1-f0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9ab7/11275690/5615512b0b4c/nihpp-2024.07.16.24310297v1-f0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9ab7/11275690/3aad29aa1dbe/nihpp-2024.07.16.24310297v1-f0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9ab7/11275690/f9e54ecb0684/nihpp-2024.07.16.24310297v1-f0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9ab7/11275690/26e2cdc8ce7f/nihpp-2024.07.16.24310297v1-f0005.jpg

相似文献

1
Evaluating AI Proficiency in Nuclear Cardiology: Large Language Models take on the Board Preparation Exam.评估人工智能在核心脏病学方面的能力:大型语言模型参加资格考试准备。
medRxiv. 2024 Jul 16:2024.07.16.24310297. doi: 10.1101/2024.07.16.24310297.
2
Evaluating AI proficiency in nuclear cardiology: Large language models take on the board preparation exam.评估人工智能在核心脏病学方面的熟练程度:大型语言模型参加资格考试。
J Nucl Cardiol. 2025 Mar;45:102089. doi: 10.1016/j.nuclcard.2024.102089. Epub 2024 Nov 29.
3
Accuracy and quality of ChatGPT-4o and Google Gemini performance on image-based neurosurgery board questions.ChatGPT-4o和谷歌Gemini在基于图像的神经外科委员会问题上的表现准确性和质量。
Neurosurg Rev. 2025 Mar 25;48(1):320. doi: 10.1007/s10143-025-03472-7.
4
Assessing the performance of Microsoft Copilot, GPT-4 and Google Gemini in ophthalmology.评估Microsoft Copilot、GPT-4和Google Gemini在眼科领域的性能。
Can J Ophthalmol. 2025 Feb 4. doi: 10.1016/j.jcjo.2025.01.001.
5
Benchmarking Vision Capabilities of Large Language Models in Surgical Examination Questions.大型语言模型在外科检查问题中的视觉能力基准测试
J Surg Educ. 2025 Apr;82(4):103442. doi: 10.1016/j.jsurg.2025.103442. Epub 2025 Feb 9.
6
Performance of GPT-4 Turbo and GPT-4o in Korean Society of Radiology In-Training Examinations.GPT-4 Turbo和GPT-4o在韩国放射学会住院医师培训考试中的表现。
Korean J Radiol. 2025 Jun;26(6):524-531. doi: 10.3348/kjr.2024.1096. Epub 2025 Apr 17.
7
Comparative Analysis of ChatGPT-4o and Gemini Advanced Performance on Diagnostic Radiology In-Training Exams.ChatGPT-4o与Gemini在放射诊断学培训考试中的性能对比分析
Cureus. 2025 Mar 20;17(3):e80874. doi: 10.7759/cureus.80874. eCollection 2025 Mar.
8
Diagnostic accuracy of vision-language models on Japanese diagnostic radiology, nuclear medicine, and interventional radiology specialty board examinations.视觉语言模型在日本放射诊断学、核医学和介入放射学专业委员会考试中的诊断准确性。
Jpn J Radiol. 2024 Dec;42(12):1392-1398. doi: 10.1007/s11604-024-01633-0. Epub 2024 Jul 20.
9
Evaluating the Effectiveness of advanced large language models in medical Knowledge: A Comparative study using Japanese national medical examination.评估先进的大型语言模型在医学知识方面的有效性:使用日本国家医学考试的比较研究。
Int J Med Inform. 2025 Jan;193:105673. doi: 10.1016/j.ijmedinf.2024.105673. Epub 2024 Oct 28.
10
Evaluating Bard Gemini Pro and GPT-4 Vision Against Student Performance in Medical Visual Question Answering: Comparative Case Study.在医学视觉问答中评估Bard Gemini Pro和GPT-4 Vision对学生表现的影响:比较案例研究
JMIR Form Res. 2024 Dec 17;8:e57592. doi: 10.2196/57592.

本文引用的文献

1
Comparing Diagnostic Accuracy of Radiologists versus GPT-4V and Gemini Pro Vision Using Image Inputs from Diagnosis Please Cases.比较放射科医生与 GPT-4V 和 Gemini Pro Vision 使用诊断请案例的图像输入的诊断准确性。
Radiology. 2024 Jul;312(1):e240273. doi: 10.1148/radiol.240273.
2
Evaluation of responses to cardiac imaging questions by the artificial intelligence large language model ChatGPT.评估人工智能大型语言模型 ChatGPT 对心脏成像问题的回答。
Clin Imaging. 2024 Aug;112:110193. doi: 10.1016/j.clinimag.2024.110193. Epub 2024 May 23.
3
GPT-4 Turbo with Vision fails to outperform text-only GPT-4 Turbo in the Japan Diagnostic Radiology Board Examination.
GPT-4 Turbo with Vision 在日本诊断放射学委员会考试中未能优于仅文本的 GPT-4 Turbo。
Jpn J Radiol. 2024 Aug;42(8):918-926. doi: 10.1007/s11604-024-01561-z. Epub 2024 May 11.
4
Performance of Google's Artificial Intelligence Chatbot "Bard" (Now "Gemini") on Ophthalmology Board Exam Practice Questions.谷歌人工智能聊天机器人“巴德”(现称“双子座”)在眼科委员会考试练习题上的表现。
Cureus. 2024 Mar 31;16(3):e57348. doi: 10.7759/cureus.57348. eCollection 2024 Mar.
5
ChatGPT performance on the American Shoulder and Elbow Surgeons maintenance of certification exam.ChatGPT 在美肩肘外科医生认证考试维护部分的表现。
J Shoulder Elbow Surg. 2024 Sep;33(9):1888-1893. doi: 10.1016/j.jse.2024.02.029. Epub 2024 Apr 4.
6
Performance of GPT-4V in Answering the Japanese Otolaryngology Board Certification Examination Questions: Evaluation Study.GPT-4V 在回答日本耳鼻喉科学委员会认证考试问题方面的表现:评估研究。
JMIR Med Educ. 2024 Mar 28;10:e57054. doi: 10.2196/57054.
7
Performance of a Large Language Model on Japanese Emergency Medicine Board Certification Examinations.大型语言模型在日本急诊医学委员会认证考试中的表现。
J Nippon Med Sch. 2024 May 21;91(2):155-161. doi: 10.1272/jnms.JNMS.2024_91-205. Epub 2024 Mar 2.
8
Performance evaluation of ChatGPT, GPT-4, and Bard on the official board examination of the Japan Radiology Society.ChatGPT、GPT-4 和 Bard 在日本放射学会官方董事会考试中的表现评估。
Jpn J Radiol. 2024 Feb;42(2):201-207. doi: 10.1007/s11604-023-01491-2. Epub 2023 Oct 4.
9
Performance of ChatGPT and GPT-4 on Neurosurgery Written Board Examinations.ChatGPT和GPT-4在神经外科笔试中的表现。
Neurosurgery. 2023 Dec 1;93(6):1353-1365. doi: 10.1227/neu.0000000000002632. Epub 2023 Aug 15.
10
Large language models in medicine.医学中的大型语言模型。
Nat Med. 2023 Aug;29(8):1930-1940. doi: 10.1038/s41591-023-02448-8. Epub 2023 Jul 17.