Department of Psychology, University of Amsterdam, Amsterdam, The Netherlands.
Psychon Bull Rev. 2019 Aug;26(4):1070-1098. doi: 10.3758/s13423-018-01563-9.
Evidence accumulations models (EAMs) have become the dominant modeling framework within rapid decision-making, using choice response time distributions to make inferences about the underlying decision process. These models are often applied to empirical data as "measurement tools", with different theoretical accounts being contrasted within the framework of the model. Some method is then needed to decide between these competing theoretical accounts, as only assessing the models on their ability to fit trends in the empirical data ignores model flexibility, and therefore, creates a bias towards more flexible models. However, there is no objectively optimal method to select between models, with methods varying in both their computational tractability and theoretical basis. I provide a systematic comparison between nine different model selection methods using a popular EAM-the linear ballistic accumulator (LBA; Brown & Heathcote, Cognitive Psychology 57(3), 153-178 2008)-in a large-scale simulation study and the empirical data of Dutilh et al. (Psychonomic Bulletin and Review, 1-19 2018). I find that the "predictive accuracy" class of methods (i.e., the Akaike Information Criterion [AIC], the Deviance Information Criterion [DIC], and the Widely Applicable Information Criterion [WAIC]) make different inferences to the "Bayes factor" class of methods (i.e., the Bayesian Information Criterion [BIC], and Bayes factors) in many, but not all, instances, and that the simpler methods (i.e., AIC and BIC) make inferences that are highly consistent with their more complex counterparts. These findings suggest that researchers should be able to use simpler "parameter counting" methods when applying the LBA and be confident in their inferences, but that researchers need to carefully consider and justify the general class of model selection method that they use, as different classes of methods often result in different inferences.
证据积累模型 (EAMs) 已成为快速决策中的主导建模框架,使用选择反应时分布来推断潜在的决策过程。这些模型通常作为“测量工具”应用于实证数据,在模型框架内对比不同的理论解释。然后,需要有一种方法来在这些竞争的理论解释之间做出选择,因为仅根据模型拟合经验数据趋势的能力来评估模型会忽略模型的灵活性,从而导致更灵活的模型产生偏差。然而,没有客观最优的方法来选择模型,方法在计算可操作性和理论基础上都有所不同。我在一项大规模模拟研究中,以及 Dutilh 等人的实证数据中,使用一种流行的 EAM——线性弹道积累器 (LBA; Brown & Heathcote, Cognitive Psychology 57(3), 153-178 2008),对九种不同的模型选择方法进行了系统比较。我发现,“预测准确性”类方法(即 Akaike 信息准则 [AIC]、偏差信息准则 [DIC] 和广泛适用信息准则 [WAIC])与“贝叶斯因子”类方法(即贝叶斯信息准则 [BIC] 和贝叶斯因子)在许多情况下做出了不同的推断,但并非所有情况下都是如此,并且更简单的方法(即 AIC 和 BIC)做出的推断与它们更复杂的对应方法高度一致。这些发现表明,研究人员在应用 LBA 时应该能够使用更简单的“参数计数”方法,并对他们的推断有信心,但研究人员需要仔细考虑和证明他们使用的模型选择方法的一般类别,因为不同类别的方法通常会导致不同的推断。