Rios Joseph A
University of Minnesota, Twin Cities, Minneapolis, MN, USA.
Educ Psychol Meas. 2022 Feb;82(1):122-150. doi: 10.1177/00131644211003640. Epub 2021 Apr 21.
The presence of rapid guessing (RG) presents a challenge to practitioners in obtaining accurate estimates of measurement properties and examinee ability. In response to this concern, researchers have utilized response times as a proxy of RG and have attempted to improve parameter estimation accuracy by filtering RG responses using popular scoring approaches, such as the effort-moderated item response theory (EM-IRT) model. However, such an approach assumes that RG can be correctly identified based on an indirect proxy of examinee behavior. A failure to meet this assumption leads to the inclusion of distortive and psychometrically uninformative information in parameter estimates. To address this issue, a simulation study was conducted to examine how violations to the assumption of correct RG classification influences EM-IRT item and ability parameter estimation accuracy and compares these results with parameter estimates from the three-parameter logistic (3PL) model, which includes RG responses in scoring. Two RG misclassification factors were manipulated: type (underclassification vs. overclassification) and rate (10%, 30%, and 50%). Results indicated that the EM-IRT model provided improved item parameter estimation over the 3PL model regardless of misclassification type and rate. Furthermore, under most conditions, increased rates of RG underclassification were associated with the greatest bias in ability parameter estimates from the EM-IRT model. In spite of this, the EM-IRT model with RG misclassifications demonstrated more accurate ability parameter estimation than the 3PL model when the mean ability of RG subgroups did not differ. This suggests that in certain situations it may be better for practitioners to (a) imperfectly identify RG than to ignore the presence of such invalid responses and (b) select liberal over conservative response time thresholds to mitigate bias from underclassified RG.
快速猜测(RG)的存在给从业者在获取测量属性和考生能力的准确估计方面带来了挑战。针对这一问题,研究人员将反应时间用作RG的替代指标,并试图通过使用流行的评分方法(如努力调节项目反应理论(EM-IRT)模型)过滤RG反应来提高参数估计的准确性。然而,这种方法假设可以基于考生行为的间接替代指标正确识别RG。如果不能满足这一假设,就会导致在参数估计中纳入扭曲且在心理测量学上无信息价值的信息。为了解决这个问题,进行了一项模拟研究,以检验违反正确RG分类假设如何影响EM-IRT项目和能力参数估计的准确性,并将这些结果与三参数逻辑斯蒂(3PL)模型的参数估计结果进行比较,3PL模型在评分时包括了RG反应。操纵了两个RG误分类因素:类型(分类不足与分类过度)和比率(10%、30%和50%)。结果表明,无论误分类类型和比率如何,EM-IRT模型在项目参数估计方面都优于3PL模型。此外,在大多数情况下,RG分类不足比率的增加与EM-IRT模型能力参数估计中最大的偏差相关。尽管如此,当RG亚组的平均能力没有差异时,存在RG误分类的EM-IRT模型在能力参数估计方面比3PL模型更准确。这表明在某些情况下,对于从业者来说,(a)不完全识别RG可能比忽略此类无效反应的存在更好,并且(b)选择宽松而非保守的反应时间阈值以减轻分类不足的RG带来的偏差可能更好。