Garabet Razmig, Mackey Brendan P, Cross James, Weingarten Michael
Drexel University College of Medicine, Philadelphia, PA USA.
Med Sci Educ. 2023 Dec 27;34(1):145-152. doi: 10.1007/s40670-023-01956-z. eCollection 2024 Feb.
We assessed the performance of OpenAI's ChatGPT-4 on United States Medical Licensing Exam STEP 1 style questions across the systems and disciplines appearing on the examination. ChatGPT-4 answered 86% of the 1300 questions accurately, exceeding the estimated passing score of 60% with no significant differences in performance across clinical domains. Findings demonstrated an improvement over earlier models as well as consistent performance in topics ranging from complex biological processes to ethical considerations in patient care. Its proficiency provides support for the use of artificial intelligence (AI) as an interactive learning tool and furthermore raises questions about how the technology can be used to educate students in the preclinical component of their medical education. The authors provide an example and discuss how students can leverage AI to receive real-time analogies and explanations tailored to their desired level of education. An appropriate application of this technology potentially enables enhancement of learning outcomes for medical students in the preclinical component of their education.
我们评估了OpenAI的ChatGPT-4在涵盖美国医师执照考试第一步中出现的各个系统和学科的题型上的表现。ChatGPT-4准确回答了1300道题中的86%,超过了估计的60%及格分数,且在各临床领域的表现无显著差异。研究结果表明,与早期模型相比有改进,并且在从复杂生物过程到患者护理中的伦理考量等一系列主题上表现稳定。其熟练程度为将人工智能(AI)用作交互式学习工具提供了支持,此外还引发了关于如何利用该技术在医学教育的临床前阶段对学生进行教育的问题。作者提供了一个示例,并讨论了学生如何利用AI获得针对其期望教育水平量身定制的实时类比和解释。这项技术的适当应用有可能提高医学生在其教育临床前阶段的学习成果。