Department of Neuroscience, University of Pennsylvania, Philadelphia, Pennsylvania, USA.
PLoS Comput Biol. 2013 Apr;9(4):e1003015. doi: 10.1371/journal.pcbi.1003015. Epub 2013 Apr 4.
Fitting models to behavior is commonly used to infer the latent computational factors responsible for generating behavior. However, the complexity of many behaviors can handicap the interpretation of such models. Here we provide perspectives on problems that can arise when interpreting parameter fits from models that provide incomplete descriptions of behavior. We illustrate these problems by fitting commonly used and neurophysiologically motivated reinforcement-learning models to simulated behavioral data sets from learning tasks. These model fits can pass a host of standard goodness-of-fit tests and other model-selection diagnostics even when the models do not provide a complete description of the behavioral data. We show that such incomplete models can be misleading by yielding biased estimates of the parameters explicitly included in the models. This problem is particularly pernicious when the neglected factors are unknown and therefore not easily identified by model comparisons and similar methods. An obvious conclusion is that a parsimonious description of behavioral data does not necessarily imply an accurate description of the underlying computations. Moreover, general goodness-of-fit measures are not a strong basis to support claims that a particular model can provide a generalized understanding of the computations that govern behavior. To help overcome these challenges, we advocate the design of tasks that provide direct reports of the computational variables of interest. Such direct reports complement model-fitting approaches by providing a more complete, albeit possibly more task-specific, representation of the factors that drive behavior. Computational models then provide a means to connect such task-specific results to a more general algorithmic understanding of the brain.
将模型拟合到行为中通常用于推断产生行为的潜在计算因素。然而,许多行为的复杂性可能会妨碍对这些模型的解释。本文提供了一些观点,讨论了当模型对行为的描述不完整时,从模型参数拟合中可能出现的问题。我们通过将常用的、具有神经生理学意义的强化学习模型拟合到学习任务的模拟行为数据集上来举例说明这些问题。即使模型没有对行为数据进行完整描述,这些模型拟合也可以通过一系列标准的拟合优度测试和其他模型选择诊断。我们表明,即使模型通过了这些测试,不完整的模型也可能会产生有偏差的参数估计,从而产生误导。当被忽略的因素未知且无法通过模型比较和类似方法轻易识别时,这个问题尤其严重。一个明显的结论是,对行为数据的简洁描述并不一定意味着对潜在计算的准确描述。此外,一般的拟合优度度量并不是支持特定模型可以提供对控制行为的计算的一般性理解的有力依据。为了帮助克服这些挑战,我们提倡设计提供感兴趣的计算变量的直接报告的任务。这种直接报告通过提供更完整的、尽管可能更特定于任务的驱动行为的因素表示,与模型拟合方法相辅相成。然后,计算模型可以将这些特定于任务的结果与对大脑的更一般算法理解联系起来。