Magruder Matthew L, Miskiewicz Michael, Rodriguez Ariel N, Ng Mitchell, Abdelgawad Amr
Department of Orthopaedic Surgery, Maimonides Medical Center, Brooklyn, NY, USA.
Renaissance School of Medicine at Stony Brook University Medical Center, Stony Brook, NY, USA.
Surgeon. 2025 Jun;23(3):187-191. doi: 10.1016/j.surge.2025.04.004. Epub 2025 Apr 22.
Recent advancements in large language model (LLM) artificial intelligence (AI) systems, like ChatGPT, have showcased ability in answering standardized examination questions, but their performance is variable. The goal of this study was to compare the performance of standard ChatGPT-4 with a custom-trained ChatGPT model taking the Orthopaedic Surgery In-Training Examination (OITE).
Practice questions for the 2022 OITE, made available on the AAOS-ResStudy website (aaos.org/education/examinations/ResStudy), were used for this study. Question stems were uploaded to both standard ChatGPT-4 and the custom-trained ChatGPT model (Orthopod), and the responses were documented as correct or incorrect. For questions containing media elements, screenshots were converted to PNG files and uploaded to ChatGPT. Evaluation of the AI's performance included descriptive statistics to determine the percent of questions answered correctly or incorrectly.
Two-hundred and seven questions were analyzed with both ChatGPT 4.0 and Orthopod. ChatGPT correctly answered 73.43 % (152/207) of the questions, while Orthopod correctly answered 71.01 % (147/207) of the questions. There was no significant difference in performance of either language model based on inclusion of media or question category.
ChatGPT 4.0 and Orthopod correctly answered 73.43 % and 71.01 % of OITE practice questions correctly. Both systems provided well-reasoned answers in response to multiple choice questions. The thoughtfully articulated responses and well-supported explanations offered by both systems may prove to be a valuable educational resource for orthopedic residents as they prepare for upcoming board-style exams.
IV.
大型语言模型(LLM)人工智能(AI)系统(如ChatGPT)的最新进展已展示出回答标准化考试问题的能力,但其表现参差不齐。本研究的目的是比较标准ChatGPT-4与经过定制训练的ChatGPT模型在参加骨外科住院医师培训考试(OITE)时的表现。
本研究使用了美国骨科学会(AAOS)ResStudy网站(aaos.org/education/examinations/ResStudy)上提供的2022年OITE练习题。问题题干被上传至标准ChatGPT-4和经过定制训练的ChatGPT模型(Orthopod),其回答被记录为正确或错误。对于包含媒体元素的问题,截图被转换为PNG文件并上传至ChatGPT。对AI表现的评估包括描述性统计,以确定回答正确或错误的问题百分比。
使用ChatGPT 4.0和Orthopod对207道问题进行了分析。ChatGPT正确回答了73.43%(152/207)的问题,而Orthopod正确回答了71.01%(147/207)的问题。基于是否包含媒体或问题类别,两种语言模型的表现均无显著差异。
ChatGPT 4.0和Orthopod分别正确回答了73.43%和71.01%的OITE练习题。两个系统针对多项选择题都给出了经过充分推理的答案。两个系统给出的条理清晰的回答和有充分依据的解释,可能会成为骨科住院医师备考即将到来的委员会式考试时的宝贵教育资源。
四级