Aarhus University, Department of Chemistry, Aarhus, Denmark.
Aarhus University, Danish Research Institute of Translational Neuroscience (DANDRITE), Aarhus, Denmark.
PLoS Comput Biol. 2021 Oct 4;17(10):e1009452. doi: 10.1371/journal.pcbi.1009452. eCollection 2021 Oct.
Choice history effects describe how future choices depend on the history of past choices. In experimental tasks this is typically framed as a bias because it often diminishes the experienced reward rates. However, in natural habitats, choices made in the past constrain choices that can be made in the future. For foraging animals, the probability of earning a reward in a given patch depends on the degree to which the animals have exploited the patch in the past. One problem with many experimental tasks that show choice history effects is that such tasks artificially decouple choice history from its consequences on reward availability over time. To circumvent this, we use a variable interval (VI) reward schedule that reinstates a more natural contingency between past choices and future reward availability. By examining the behavior of optimal agents in the VI task we discover that choice history effects observed in animals serve to maximize reward harvesting efficiency. We further distil the function of choice history effects by manipulating first- and second-order statistics of the environment. We find that choice history effects primarily reflect the growth rate of the reward probability of the unchosen option, whereas reward history effects primarily reflect environmental volatility. Based on observed choice history effects in animals, we develop a reinforcement learning model that explicitly incorporates choice history over multiple time scales into the decision process, and we assess its predictive adequacy in accounting for the associated behavior. We show that this new variant, known as the double trace model, has a higher performance in predicting choice data, and shows near optimal reward harvesting efficiency in simulated environments. These results suggests that choice history effects may be adaptive for natural contingencies between consumption and reward availability. This concept lends credence to a normative account of choice history effects that extends beyond its description as a bias.
选择历史效应描述了未来的选择如何取决于过去选择的历史。在实验任务中,这通常被描述为一种偏差,因为它往往会降低体验到的奖励率。然而,在自然栖息地中,过去做出的选择会限制未来可以做出的选择。对于觅食动物来说,在给定的斑块中获得奖励的概率取决于动物过去对斑块的开发程度。许多显示选择历史效应的实验任务存在的一个问题是,这些任务将选择历史与其对时间内奖励可用性的影响人为地分开。为了解决这个问题,我们使用可变间隔(VI)奖励时间表,在过去的选择和未来的奖励可用性之间恢复更自然的联系。通过研究 VI 任务中最优代理的行为,我们发现动物中观察到的选择历史效应有助于最大化奖励收获效率。我们通过操纵环境的一阶和二阶统计数据进一步提炼选择历史效应的功能。我们发现,选择历史效应主要反映了未选中选项的奖励概率的增长率,而奖励历史效应主要反映了环境的波动性。基于在动物中观察到的选择历史效应,我们开发了一种强化学习模型,该模型将多个时间尺度的选择历史明确纳入决策过程,并评估其在解释相关行为方面的预测充分性。我们表明,这种新的变体,称为双痕迹模型,在预测选择数据方面表现出更高的性能,并在模拟环境中显示出接近最优的奖励收获效率。这些结果表明,选择历史效应可能适应于消费和奖励可用性之间的自然联系。这一概念为选择历史效应的规范性解释提供了依据,超越了其作为一种偏差的描述。