Huang Qi Helen, Bolt Daniel M
University of Wisconsin-Madison, USA.
Educ Psychol Meas. 2023 Aug;83(4):808-830. doi: 10.1177/00131644221117194. Epub 2022 Aug 18.
Previous studies have demonstrated evidence of latent skill continuity even in tests intentionally designed for measurement of binary skills. In addition, the assumption of binary skills when continuity is present has been shown to potentially create a lack of invariance in item and latent ability parameters that may undermine applications. In this article, we examine measurement of growth as one such application, and consider multidimensional item response theory (MIRT) as a competing alternative. Motivated by prior findings concerning the effects of skill continuity, we study the relative robustness of cognitive diagnostic models (CDMs) and (M)IRT models in the measurement of growth under both binary and continuous latent skill distributions. We find CDMs to be a less robust way of quantifying growth under misspecification, and subsequently provide a real-data example suggesting underestimation of growth as a likely consequence. It is suggested that researchers should regularly attend to the assumptions associated with the use of latent binary skills and consider (M)IRT as a potentially more robust alternative if unsure of their discrete nature.
以往的研究已经证明,即使在故意设计用于测量二元技能的测试中,也存在潜在技能连续性的证据。此外,当存在连续性时,二元技能的假设已被证明可能会导致项目和潜在能力参数缺乏不变性,这可能会破坏应用。在本文中,我们将检验作为此类应用之一的增长测量,并将多维项目反应理论(MIRT)视为一种竞争性替代方法。受先前关于技能连续性影响的研究结果的启发,我们研究了认知诊断模型(CDMs)和(M)IRT模型在二元和连续潜在技能分布下测量增长时的相对稳健性。我们发现,在错误设定的情况下,CDMs是一种量化增长的稳健性较差的方法,随后提供了一个实际数据示例,表明增长被低估可能是一个后果。建议研究人员应经常关注与使用潜在二元技能相关的假设,如果不确定其离散性质,可将(M)IRT视为一种可能更稳健的替代方法。