Gebremichael Mengistu Welday, Baraki Beyene, Mehari Mihret-Ab, Assalfew Betel
Mekelle University College of Health Sciences, P.O. Box: 1871, Mekelle City, Tigray, Ethiopia.
Department of Pedagogy, Mekelle University College of Behavioral Sciences, Mekelle City, Ethiopia.
BMC Med Educ. 2025 Mar 26;25(1):441. doi: 10.1186/s12909-025-06904-6.
Rapid expansion of higher education and graduation rate in Ethiopia, doesn't tell much about quality of core processes like teaching and learning. Despite the fact that preparation of a good item is essential, little attention has been paid to examining characteristics of exam items such as difficulty, discrimination, and distractor efficiency. Unless these characters are examined results may lead to biased evaluation. To address the current issue, this study aimed to carry out item analysis of multiple choice questions (MCQs) in undergraduate health sciences students of Mekelle University, Tigray, Ethiopia in order to evaluate correlations between item difficulty, item discrimination and distraction efficiency.
Institution based cross sectional study was conducted from March to April 2020. A total of 189 exam papers with 539 multiple choice items and 2224 distracters were analyzed in terms of difficulty, discrimination, distractor efficiency and internal consistency reliability. SPSS version 25 was used for analysis & findings presented using tables & figures. Pearson's correlation & one-way analysis of variance were used.
In this study mean difficulty ranged from 0.57-0.81, and discrimination from 0.27-0.52. Difficulty and discrimination were negatively correlated. Six departments achieved excellent reliability, indicating r>0.9. However, the reliability score of Medicine and Pharmacy departments was poor (r<0.5), suggesting areas for improvement. The remaining two departments exhibited acceptable reliability. More than three fourth of the items had non-functional distractors.
The internal consistency of the exams from six departments demonstrated excellent reliability. However, the exams from the Medicine and Pharmacy departments showed areas for improvement. In order to enhance assessment effectiveness, it's crucial to incorporate items that exhibit both desirable difficulty and high discrimination along with functionally realistic distracters. This approach ensures assessments accuracy.
埃塞俄比亚高等教育的迅速扩张和毕业率,并未充分反映教学等核心过程的质量。尽管编制优质试题至关重要,但对于试题的难度、区分度和干扰项有效性等特征的考察却很少。除非对这些特征进行考察,否则结果可能导致有偏差的评估。为解决当前这一问题,本研究旨在对埃塞俄比亚提格雷州梅克内斯大学本科健康科学专业学生的多项选择题进行试题分析,以评估试题难度、区分度和干扰项有效性之间的相关性。
2020年3月至4月进行了基于机构的横断面研究。共分析了189份试卷,其中包含539道多项选择题和2224个干扰项,分析内容包括难度、区分度、干扰项有效性和内部一致性信度。使用SPSS 25版进行分析,并以表格和图表形式呈现结果。采用皮尔逊相关性分析和单因素方差分析。
本研究中,平均难度范围为0.57 - 0.81,区分度范围为0.27 - 0.52。难度和区分度呈负相关。六个系的信度极佳,表明r>0.9。然而,医学系和药学系的信度得分较差(r<0.5),表明有待改进。其余两个系的信度可接受。超过四分之三的试题具有无效干扰项。
六个系考试的内部一致性显示出极佳的信度。然而,医学系和药学系的考试显示出有待改进的方面。为提高评估效果,纳入具有理想难度和高区分度的试题以及功能上合理的干扰项至关重要。这种方法可确保评估的准确性。