• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

与人工智能模型相比,风湿病住院医师的临床推理与知识评估:一项试点研究。

Clinical Reasoning and Knowledge Assessment of Rheumatology Residents Compared to AI Models: A Pilot Study.

作者信息

Kayacan Erdoğan Esra, Babaoğlu Hakan

机构信息

Department of Rheumatology, Division of Internal Medicine, Ankara Bilkent City Hospital, Ankara 06800, Turkey.

出版信息

J Clin Med. 2024 Dec 5;13(23):7405. doi: 10.3390/jcm13237405.

DOI:10.3390/jcm13237405
PMID:39685863
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11642710/
Abstract

The integration of artificial intelligence (AI) in medicine has progressed from rule-based systems to advanced models and is showing potential in clinical decision-making. In this study, the psychological impact of AI collaboration in clinical practice is assessed, highlighting its role as a support tool for medical residents. This study aimed to compare clinical decision-making approaches of junior rheumatology residents with both trained and untrained AI models in clinical reasoning, pre-diagnosis, first-line, and second-line management stages. Ten junior rheumatology residents and two GPT-4 models (trained and untrained) responded to 10 clinical cases, encompassing diagnostic and treatment challenges in inflammatory arthritis. The cases were evaluated using the Revised-IDEA (R-IDEA) scoring system and additional case management metrics. In addition to scoring clinical case performance, residents' attitudes toward AI integration in clinical practice were assessed through a structured questionnaire, focusing on perceptions of AI's potential after reviewing the trained GPT-4's answers. Trained GPT-4 outperformed residents across all stages, achieving significantly higher median R-IDEA scores and superior performance in pre-diagnosis, first-line, and second-line management phases. Residents expressed a positive attitude toward AI integration, with 60% favoring AI as a supportive tool in clinical practice, anticipating benefits in competence, fatigue, and burnout. Trained GPT-4 models outperform junior residents in clinical reasoning and management of rheumatology cases. Residents' positive attitudes toward AI suggest its potential as a supportive tool to enhance confidence and reduce uncertainty in clinical practice. Trained GPT-4 may be used as a supplementary tool during the early years of residency.

摘要

人工智能(AI)在医学领域的整合已从基于规则的系统发展到先进模型,并在临床决策中显示出潜力。在本研究中,评估了AI协作在临床实践中的心理影响,突出了其作为医学住院医师支持工具的作用。本研究旨在比较初级风湿病住院医师在临床推理、预诊断、一线和二线管理阶段与经过训练和未经训练的AI模型的临床决策方法。十名初级风湿病住院医师和两个GPT-4模型(经过训练和未经训练)对10个临床病例做出了回应,这些病例涵盖了炎症性关节炎的诊断和治疗挑战。使用修订后的IDEA(R-IDEA)评分系统和其他病例管理指标对病例进行评估。除了对临床病例表现进行评分外,还通过一份结构化问卷评估了住院医师对AI整合到临床实践中的态度,重点是在查看经过训练的GPT-4的答案后对AI潜力的看法。经过训练的GPT-4在所有阶段的表现均优于住院医师,在预诊断、一线和二线管理阶段的R-IDEA中位数得分显著更高,表现更优。住院医师对AI整合表达了积极态度,60%的人赞成在临床实践中使用AI作为支持工具,预计在能力、疲劳和职业倦怠方面会有好处。经过训练的GPT-4模型在风湿病病例的临床推理和管理方面优于初级住院医师。住院医师对AI的积极态度表明其有潜力作为一种支持工具,增强临床实践中的信心并减少不确定性。经过训练的GPT-4可在住院医师培训的早期用作辅助工具。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f740/11642710/a74f3bcebaf5/jcm-13-07405-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f740/11642710/1099b4c36e5a/jcm-13-07405-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f740/11642710/a74f3bcebaf5/jcm-13-07405-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f740/11642710/1099b4c36e5a/jcm-13-07405-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f740/11642710/a74f3bcebaf5/jcm-13-07405-g002.jpg

相似文献

1
Clinical Reasoning and Knowledge Assessment of Rheumatology Residents Compared to AI Models: A Pilot Study.与人工智能模型相比,风湿病住院医师的临床推理与知识评估:一项试点研究。
J Clin Med. 2024 Dec 5;13(23):7405. doi: 10.3390/jcm13237405.
2
Gemini AI vs. ChatGPT: A comprehensive examination alongside ophthalmology residents in medical knowledge.Gemini人工智能与ChatGPT对比:与眼科住院医师一起对医学知识进行的全面考察
Graefes Arch Clin Exp Ophthalmol. 2025 Feb;263(2):527-536. doi: 10.1007/s00417-024-06625-4. Epub 2024 Sep 15.
3
GPT-4 Artificial Intelligence Model Outperforms ChatGPT, Medical Students, and Neurosurgery Residents on Neurosurgery Written Board-Like Questions.GPT-4人工智能模型在类似神经外科书面考试的问题上表现优于ChatGPT、医学生和神经外科住院医师。
World Neurosurg. 2023 Nov;179:e160-e165. doi: 10.1016/j.wneu.2023.08.042. Epub 2023 Aug 18.
4
Performance Comparison of ChatGPT-4 and Japanese Medical Residents in the General Medicine In-Training Examination: Comparison Study.ChatGPT-4与日本内科住院医师在普通内科培训考试中的表现比较:比较研究
JMIR Med Educ. 2023 Dec 6;9:e52202. doi: 10.2196/52202.
5
Evaluating ChatGPT o1's Capabilities in Peripheral Nerve Surgery: Advancing Artificial Intelligence in Clinical Practice.评估ChatGPT-1在外周神经外科手术中的能力:推动人工智能在临床实践中的应用。
World Neurosurg. 2025 Apr;196:123753. doi: 10.1016/j.wneu.2025.123753. Epub 2025 Mar 6.
6
The future of AI clinicians: assessing the modern standard of chatbots and their approach to diagnostic uncertainty.人工智能临床医生的未来:评估现代聊天机器人的标准及其对诊断不确定性的处理方法。
BMC Med Educ. 2024 Oct 11;24(1):1133. doi: 10.1186/s12909-024-06115-5.
7
Diagnostic accuracy of large language models in psychiatry.精神科大语言模型的诊断准确性。
Asian J Psychiatr. 2024 Oct;100:104168. doi: 10.1016/j.ajp.2024.104168. Epub 2024 Jul 25.
8
Assessing the role of GPT-4 in thyroid ultrasound diagnosis and treatment recommendations: enhancing interpretability with a chain of thought approach.评估GPT-4在甲状腺超声诊断及治疗建议中的作用:采用思维链方法提高可解释性
Quant Imaging Med Surg. 2024 Feb 1;14(2):1602-1615. doi: 10.21037/qims-23-1180. Epub 2024 Jan 11.
9
The Role of Artificial Intelligence in Surgery: What do General Surgery Residents Think?人工智能在外科手术中的作用:普通外科住院医师怎么看?
Am Surg. 2024 Apr;90(4):541-549. doi: 10.1177/00031348231209524. Epub 2023 Oct 20.
10
Comparative evaluation of artificial intelligence models GPT-4 and GPT-3.5 in clinical decision-making in sports surgery and physiotherapy: a cross-sectional study.人工智能模型GPT-4和GPT-3.5在运动外科和物理治疗临床决策中的比较评估:一项横断面研究。
BMC Med Inform Decis Mak. 2025 Apr 14;25(1):163. doi: 10.1186/s12911-025-02996-8.

本文引用的文献

1
Prediction of a Cephalometric Parameter and Skeletal Patterns from Lateral Profile Photographs: A Retrospective Comparative Analysis of Regression Convolutional Neural Networks.基于侧位照片预测头影测量参数和骨骼类型:回归卷积神经网络的回顾性对比分析
J Clin Med. 2024 Oct 23;13(21):6346. doi: 10.3390/jcm13216346.
2
Assessing the Role of the Generative Pretrained Transformer (GPT) in Alzheimer's Disease Management: Comparative Study of Neurologist- and Artificial Intelligence-Generated Responses.评估生成式预训练转换器(GPT)在阿尔茨海默病管理中的作用:神经科医生和人工智能生成的回复的对比研究。
J Med Internet Res. 2024 Oct 31;26:e51095. doi: 10.2196/51095.
3
Machine Learning Algorithm-Aided Determination of Predictors of Mortality from Diabetic Foot Sepsis at a Regional Hospital in South Africa During the COVID-19 Pandemic.
机器学习算法辅助确定南非一家地区医院在 COVID-19 大流行期间糖尿病足败血症死亡率的预测因素。
Medicina (Kaunas). 2024 Oct 20;60(10):1718. doi: 10.3390/medicina60101718.
4
An ethics assessment tool for artificial intelligence implementation in healthcare: CARE-AI.一种用于医疗保健领域人工智能实施的伦理评估工具:CARE-AI。
Nat Med. 2024 Nov;30(11):3038-3039. doi: 10.1038/s41591-024-03310-1.
5
Clinical Reasoning of a Generative Artificial Intelligence Model Compared With Physicians.生成式人工智能模型与医生的临床推理比较
JAMA Intern Med. 2024 May 1;184(5):581-583. doi: 10.1001/jamainternmed.2024.0295.
6
ChatGPT versus clinician: challenging the diagnostic capabilities of artificial intelligence in dermatology.ChatGPT 与临床医生:人工智能在皮肤病学诊断能力方面的挑战。
Clin Exp Dermatol. 2024 Jun 25;49(7):707-710. doi: 10.1093/ced/llad402.
7
A pilot study on the efficacy of GPT-4 in providing orthopedic treatment recommendations from MRI reports.GPT-4 在提供 MRI 报告中的骨科治疗建议方面的功效的初步研究。
Sci Rep. 2023 Nov 17;13(1):20159. doi: 10.1038/s41598-023-47500-2.
8
Comparing ChatGPT and GPT-4 performance in USMLE soft skill assessments.比较 ChatGPT 和 GPT-4 在 USMLE 软技能评估中的表现。
Sci Rep. 2023 Oct 1;13(1):16492. doi: 10.1038/s41598-023-43436-9.
9
Can Chat-GPT a substitute for urological resident physician in diagnosing diseases?: a preliminary conclusion from an exploratory investigation.Chat-GPT能否替代泌尿外科住院医师进行疾病诊断?:一项探索性调查的初步结论。
World J Urol. 2023 Sep;41(9):2569-2571. doi: 10.1007/s00345-023-04539-0. Epub 2023 Jul 28.
10
How Does ChatGPT Perform on the United States Medical Licensing Examination (USMLE)? The Implications of Large Language Models for Medical Education and Knowledge Assessment.ChatGPT在美国医师执照考试(USMLE)中的表现如何?大语言模型对医学教育和知识评估的影响。
JMIR Med Educ. 2023 Feb 8;9:e45312. doi: 10.2196/45312.