• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

大语言模型有助于解决放射学研究中所需的生物统计学和编码问题。

Large Language Models can Help with Biostatistics and Coding Needed in Radiology Research.

作者信息

Ghosh Adarsh, Li Hailong, Trout Andrew T

机构信息

Department of Radiology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio (A.G., H.L., A.T.T.); Department of Radiology, Nationwide Children's Hospital, Columbus, Ohio (A.G.).

Department of Radiology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio (A.G., H.L., A.T.T.); Department of Radiology, University of Cincinnati College of Medicine, Cincinnati, Ohio (H.L., A.T.T.); Imaging Research Center, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio (H.L.).

出版信息

Acad Radiol. 2025 Feb;32(2):604-611. doi: 10.1016/j.acra.2024.09.042. Epub 2024 Oct 15.

DOI:10.1016/j.acra.2024.09.042
PMID:39406582
Abstract

INTRODUCTION

Original research in radiology often involves handling large datasets, data manipulation, statistical tests, and coding. Recent studies show that large language models (LLMs) can solve bioinformatics tasks, suggesting their potential in radiology research. This study evaluates an LLM's ability to provide statistical and deep learning solutions and code for radiology research.

MATERIALS AND METHODS

We used web-based chat interfaces available for ChatGPT-4o, ChatGPT-3.5, and Google Gemini. EXPERIMENT 1: BIOSTATISTICS AND DATA VISUALIZATION: We assessed each LLMs' ability to suggest biostatistical tests and generate R code for the same using a Cancer Imaging Archive dataset. Prompts were based on statistical analyses from a peer-reviewed manuscript. The generated code was tested in R Studio for correctness, runtime errors and the ability to generate the requested visualization. EXPERIMENT 2: DEEP LEARNING: We used the RSNA-STR Pneumonia Detection Challenge dataset to evaluate ChatGPT-4o and Gemini's ability to generate Python code for transformer-based image classification models (Vision Transformer ViT-B/16). The generated code was tested in a Jupiter Notebook for functionality and run time errors.

RESULTS

Out of the 8 statistical questions posed, correct statistical answers were suggested for 7 (ChatGPT-4o), 6 (ChatGPT-3.5), and 5 (Gemini) scenarios. The R code output by ChatGPT-4o had fewer runtime errors (6 out of the 7 total codes provided) compared to ChatGPT-3.5 (5/7) and Gemini (5/7). Both ChatGPT4o and Gemini were able to generate visualization requested with a few run time errors. Iteratively copying runtime errors from the code generated by ChatGPT4o into the chat helped resolve them. Gemini initially hallucinated during code generation but was able to provide accurate code on restarting the experiment. ChatGPT4-o and Gemini successfully generated initial Python code for deep learning tasks. Errors encountered during implementation were resolved through iterations using the chat interface, demonstrating LLM utility in providing baseline code for further code refinement and resolving run time errors.

CONCLUSION

LLMs can assist in coding tasks for radiology research, providing initial code for data visualization, statistical tests, and deep learning models helping researchers with foundational biostatistical knowledge. While LLM can offer a useful starting point, they require users to refine and validate the code and caution is necessary due to potential errors, the risk of hallucinations and data privacy regulations.

SUMMARY STATEMENT

LLMs can help with coding and statistical problems in radiology research. This can help primary authors trouble shoot coding needed in radiology research.

摘要

引言

放射学领域的原创研究通常涉及处理大型数据集、数据操作、统计测试和编码。最近的研究表明,大语言模型(LLMs)可以解决生物信息学任务,这暗示了它们在放射学研究中的潜力。本研究评估了一个大语言模型为放射学研究提供统计和深度学习解决方案及代码的能力。

材料与方法

我们使用了可用于ChatGPT-4o、ChatGPT-3.5和谷歌Gemini的基于网络的聊天界面。实验1:生物统计学与数据可视化:我们使用癌症影像存档数据集评估了每个大语言模型建议生物统计学测试并生成R代码的能力。提示基于一篇同行评议手稿中的统计分析。在R Studio中测试生成的代码的正确性、运行时错误以及生成所需可视化的能力。实验2:深度学习:我们使用RSNA-STR肺炎检测挑战赛数据集评估ChatGPT-4o和Gemini生成基于Transformer的图像分类模型(视觉Transformer ViT-B/16)的Python代码的能力。在Jupiter Notebook中测试生成的代码的功能和运行时错误。

结果

在提出的8个统计问题中,ChatGPT-4o针对7个、ChatGPT-3.5针对6个、Gemini针对5个场景给出了正确的统计答案。与ChatGPT-3.5(7个代码中有5个)和Gemini(7个代码中有5个)相比,ChatGPT-4o输出的R代码运行时错误更少(7个代码中有6个)。ChatGPT4o和Gemini都能够生成所需的可视化,仅有一些运行时错误。将ChatGPT4o生成的代码中的运行时错误迭代复制到聊天中有助于解决这些问题。Gemini在代码生成过程中最初出现了幻觉,但在重新启动实验后能够提供准确的代码。ChatGPT4-o和Gemini成功地为深度学习任务生成了初始Python代码。通过使用聊天界面进行迭代解决了实现过程中遇到的错误,这证明了大语言模型在提供基线代码以进行进一步代码优化和解决运行时错误方面的实用性。

结论

大语言模型可以协助放射学研究的编码任务,为数据可视化、统计测试和深度学习模型提供初始代码,帮助有基础生物统计学知识的研究人员。虽然大语言模型可以提供一个有用的起点,但它们要求用户对代码进行优化和验证,并且由于潜在的错误、幻觉风险和数据隐私法规,必须谨慎使用。

总结声明

大语言模型可以帮助解决放射学研究中的编码和统计问题。这可以帮助第一作者解决放射学研究中所需的编码问题。

相似文献

1
Large Language Models can Help with Biostatistics and Coding Needed in Radiology Research.大语言模型有助于解决放射学研究中所需的生物统计学和编码问题。
Acad Radiol. 2025 Feb;32(2):604-611. doi: 10.1016/j.acra.2024.09.042. Epub 2024 Oct 15.
2
Comparative Analysis of ChatGPT-4o and Gemini Advanced Performance on Diagnostic Radiology In-Training Exams.ChatGPT-4o与Gemini在放射诊断学培训考试中的性能对比分析
Cureus. 2025 Mar 20;17(3):e80874. doi: 10.7759/cureus.80874. eCollection 2025 Mar.
3
Evaluating text and visual diagnostic capabilities of large language models on questions related to the Breast Imaging Reporting and Data System Atlas 5 edition.评估大语言模型在与《乳腺影像报告和数据系统》第5版相关问题上的文本和视觉诊断能力。
Diagn Interv Radiol. 2025 Mar 3;31(2):111-129. doi: 10.4274/dir.2024.242876. Epub 2024 Sep 9.
4
Comparison of ChatGPT-4o, Google Gemini 1.5 Pro, Microsoft Copilot Pro, and Ophthalmologists in the management of uveitis and ocular inflammation: A comparative study of large language models.ChatGPT-4o、谷歌Gemini 1.5 Pro、微软Copilot Pro与眼科医生在葡萄膜炎和眼部炎症管理中的比较:大型语言模型的对比研究
J Fr Ophtalmol. 2025 Apr;48(4):104468. doi: 10.1016/j.jfo.2025.104468. Epub 2025 Mar 13.
5
Accuracy and quality of ChatGPT-4o and Google Gemini performance on image-based neurosurgery board questions.ChatGPT-4o和谷歌Gemini在基于图像的神经外科委员会问题上的表现准确性和质量。
Neurosurg Rev. 2025 Mar 25;48(1):320. doi: 10.1007/s10143-025-03472-7.
6
Assessment of decision-making with locally run and web-based large language models versus human board recommendations in otorhinolaryngology, head and neck surgery.在耳鼻喉科、头颈外科中,评估本地运行和基于网络的大语言模型与人类委员会建议的决策情况。
Eur Arch Otorhinolaryngol. 2025 Mar;282(3):1593-1607. doi: 10.1007/s00405-024-09153-3. Epub 2025 Jan 10.
7
Evaluation of ChatGPT and Gemini large language models for pharmacometrics with NONMEM.评估 ChatGPT 和 Gemini 大型语言模型在 NONMEM 中的药物代谢动力学应用。
J Pharmacokinet Pharmacodyn. 2024 Jun;51(3):187-197. doi: 10.1007/s10928-024-09921-y. Epub 2024 Apr 24.
8
Comparative analysis of ChatGPT-4o mini, ChatGPT-4o and Gemini Advanced in the treatment of postmenopausal osteoporosis.ChatGPT-4o mini、ChatGPT-4o与Gemini Advanced在绝经后骨质疏松症治疗中的对比分析。
BMC Musculoskelet Disord. 2025 Apr 16;26(1):369. doi: 10.1186/s12891-025-08601-3.
9
Performance of Large Language Models ChatGPT and Gemini on Workplace Management Questions in Radiology.大型语言模型ChatGPT和Gemini在放射学工作场所管理问题上的表现
Diagnostics (Basel). 2025 Feb 19;15(4):497. doi: 10.3390/diagnostics15040497.
10
Comparative performance of artificial intelligence models in rheumatology board-level questions: evaluating Google Gemini and ChatGPT-4o.人工智能模型在风湿病委员会级问题中的比较性能:评估 Google Gemini 和 ChatGPT-4o。
Clin Rheumatol. 2024 Nov;43(11):3507-3513. doi: 10.1007/s10067-024-07154-5. Epub 2024 Sep 28.

引用本文的文献

1
The Accuracy of ChatGPT-4o in Interpreting Chest and Abdominal X-Ray Images.ChatGPT-4o 在解读胸部和腹部 X 光图像方面的准确性。
J Pers Med. 2025 May 10;15(5):194. doi: 10.3390/jpm15050194.
2
Steering veridical large language model analyses by correcting and enriching generated database queries: first steps toward ChatGPT bioinformatics.通过纠正和丰富生成的数据库查询来引导真实的大语言模型分析:迈向ChatGPT生物信息学的第一步。
Brief Bioinform. 2024 Nov 22;26(1). doi: 10.1093/bib/bbaf045.