Eleragi A M S, Miskeen Elhadi, Hussein Kamal, Rezigalla Assad Ali, Adam Masoud I E, Al-Faifi Jaber Ahmed, Alhalafi Abdullah, Al Ameer Ahmed Y, Mohammed Osama A
Department of Microbiology College of Medicine, University of Bisha, Bisha, 61922, Saudi Arabia.
Department of Obstetrics and Gynecology, College of Medicine, University of Bisha, Bisha, 61922, Saudi Arabia.
BMC Med Educ. 2025 Feb 13;25(1):233. doi: 10.1186/s12909-025-06700-2.
Assessment is a central tool that drives and shapes students learning. Multiple choice questions (MCQs) are central in medical education assessment because they evaluate knowledge across large cohorts. Good quality items will help to achieve the learning objectives and provide trustful results. This study aims to evaluate the quality of MCQs utilized in the final exams of the Principal of Diseases (PRD) course over three academic years at the College of Medicine at The University of Bisha, Saudi Arabia.
This cross-sectional institutional-based study used the final exams from the PRD course for the academic years 2016-2019. It was conducted at the College of Medicine, University of Bisha (UBCOM), Saudi Arabia (SA). The analysis process used item analysis (IA) of the PRD final theoretical examinations of the 2016-2017, 2017-2018, and 2018-2019 academic years. 80, 70, and 60 MCQ items were used per test in the above-mentioned years, respectively (210 total). The IA targets the reliability (KR20), difficulty index (DIF), discrimination index (DI), and distractor effectiveness (DE). The generated data were analyzed using SPSS (version 25.0), and statistical significance was set at P < 0.05.
The exams included 210 items. The reliability (KR20) ranged from 0.804 to 0.906. The DI indicated that 56.7% of items were excellent, 20.9% were good, 13.8% were poor, and 8.6% were defective. The DIF showed that 50.5% of items had acceptable difficulty, 37.6% were easy, and 11.9% were difficult. DE analysis revealed that 70.2% of distractors were functional, with a significant correlation between DI, DIF, and DE (P < 0.05).
Most of the examined items exhibited excellent discrimination and acceptable difficulty, with 70.2% having functional distractors, categorizing them as high-quality and well-constructed items. The study accentuates the importance of continuous item analysis to maintain and improve the quality of assessment tools used in medical education.
评估是推动和塑造学生学习的核心工具。多项选择题(MCQs)在医学教育评估中至关重要,因为它们能评估大量学生的知识水平。高质量的题目有助于实现学习目标并提供可靠的结果。本研究旨在评估沙特阿拉伯比沙大学医学院在三个学年的疾病学原理(PRD)课程期末考试中所使用的多项选择题的质量。
这项基于机构的横断面研究使用了2016 - 2019学年PRD课程的期末考试。研究在沙特阿拉伯比沙大学医学院(UBCOM)进行。分析过程采用了对2016 - 2017、2017 - 2018和2018 - 2019学年PRD期末考试的题目分析(IA)。上述年份每次考试分别使用了80、70和60道多项选择题(共210道)。题目分析的指标包括信度(KR20)、难度指数(DIF)、区分度指数(DI)和干扰项有效性(DE)。使用SPSS(25.0版)对生成的数据进行分析,设定统计学显著性为P < 0.05。
考试包括210道题目。信度(KR20)范围为0.804至0.906。区分度指数表明,56.7%的题目优秀,20.9%的题目良好,13.8%的题目较差,8.6%的题目有缺陷。难度指数显示,50.5%的题目难度可接受,37.6%的题目容易,11.9%的题目困难。干扰项有效性分析表明,70.2%的干扰项起作用,区分度指数、难度指数和干扰项有效性之间存在显著相关性(P < 0.05)。
大多数被考查的题目表现出优秀的区分度和可接受的难度,70.2%的题目有起作用的干扰项,可将它们归类为高质量且结构良好的题目。该研究强调了持续进行题目分析对于维持和提高医学教育中评估工具质量的重要性。