• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

在执行试错问题解决任务的非人类灵长类动物中工作记忆与强化学习的适应性协调

Adaptive coordination of working-memory and reinforcement learning in non-human primates performing a trial-and-error problem solving task.

作者信息

Viejo Guillaume, Girard Benoît, Procyk Emmanuel, Khamassi Mehdi

机构信息

Sorbonne Universités, UPMC Univ Paris 06, CNRS, Institute of Intelligent Systems and Robotics (ISIR), F-75005 Paris, France; Montreal Neurological Institute and Hospital, 3801 University Street, Montreal, Quebec, Canada.

Sorbonne Universités, UPMC Univ Paris 06, CNRS, Institute of Intelligent Systems and Robotics (ISIR), F-75005 Paris, France.

出版信息

Behav Brain Res. 2018 Dec 14;355:76-89. doi: 10.1016/j.bbr.2017.09.030. Epub 2017 Oct 20.

DOI:10.1016/j.bbr.2017.09.030
PMID:29061387
Abstract

Accumulating evidence suggest that human behavior in trial-and-error learning tasks based on decisions between discrete actions may involve a combination of reinforcement learning (RL) and working-memory (WM). While the understanding of brain activity at stake in this type of tasks often involve the comparison with non-human primate neurophysiological results, it is not clear whether monkeys use similar combined RL and WM processes to solve these tasks. Here we analyzed the behavior of five monkeys with computational models combining RL and WM. Our model-based analysis approach enables to not only fit trial-by-trial choices but also transient slowdowns in reaction times, indicative of WM use. We found that the behavior of the five monkeys was better explained in terms of a combination of RL and WM despite inter-individual differences. The same coordination dynamics we used in a previous study in humans best explained the behavior of some monkeys while the behavior of others showed the opposite pattern, revealing a possible different dynamics of WM process. We further analyzed different variants of the tested models to open a discussion on how the long pretraining in these tasks may have favored particular coordination dynamics between RL and WM. This points towards either inter-species differences or protocol differences which could be further tested in humans.

摘要

越来越多的证据表明,人类在基于离散动作之间的决策进行试错学习任务时的行为可能涉及强化学习(RL)和工作记忆(WM)的结合。虽然对这类任务中大脑活动的理解通常涉及与非人类灵长类动物神经生理学结果的比较,但尚不清楚猴子是否使用类似的RL和WM组合过程来解决这些任务。在这里,我们用结合RL和WM的计算模型分析了五只猴子的行为。我们基于模型的分析方法不仅能够拟合逐次试验的选择,还能拟合反应时间的短暂减慢,这表明使用了WM。我们发现,尽管存在个体差异,但五只猴子的行为用RL和WM的组合来解释更好。我们在之前一项针对人类的研究中使用的相同协调动态最能解释一些猴子的行为,而其他猴子的行为则呈现相反模式,揭示了WM过程可能存在不同的动态。我们进一步分析了测试模型的不同变体,以开启关于这些任务中的长期预训练如何可能有利于RL和WM之间特定协调动态的讨论。这指向了种间差异或实验方案差异,可在人类中进一步测试。

相似文献

1
Adaptive coordination of working-memory and reinforcement learning in non-human primates performing a trial-and-error problem solving task.在执行试错问题解决任务的非人类灵长类动物中工作记忆与强化学习的适应性协调
Behav Brain Res. 2018 Dec 14;355:76-89. doi: 10.1016/j.bbr.2017.09.030. Epub 2017 Oct 20.
2
Interactions Among Working Memory, Reinforcement Learning, and Effort in Value-Based Choice: A New Paradigm and Selective Deficits in Schizophrenia.基于价值的选择中工作记忆、强化学习和努力之间的相互作用:一种新范式及精神分裂症中的选择性缺陷。
Biol Psychiatry. 2017 Sep 15;82(6):431-439. doi: 10.1016/j.biopsych.2017.05.017. Epub 2017 May 31.
3
Working Memory Load Strengthens Reward Prediction Errors.工作记忆负荷增强奖励预测误差。
J Neurosci. 2017 Apr 19;37(16):4332-4342. doi: 10.1523/JNEUROSCI.2700-16.2017. Epub 2017 Mar 20.
4
Within- and across-trial dynamics of human EEG reveal cooperative interplay between reinforcement learning and working memory.人类脑电图的试验内和试验间动态揭示了强化学习和工作记忆之间的协同相互作用。
Proc Natl Acad Sci U S A. 2018 Mar 6;115(10):2502-2507. doi: 10.1073/pnas.1720963115. Epub 2018 Feb 20.
5
Modeling of autonomous problem solving process by dynamic construction of task models in multiple tasks environment.在多任务环境中通过动态构建任务模型对自主问题解决过程进行建模。
Neural Netw. 2006 Oct;19(8):1169-80. doi: 10.1016/j.neunet.2006.05.037. Epub 2006 Sep 20.
6
Neural Index of Reinforcement Learning Predicts Improved Stimulus-Response Retention under High Working Memory Load.神经强化学习指数预测在高工作记忆负荷下改善刺激-反应保持。
J Neurosci. 2023 Apr 26;43(17):3131-3143. doi: 10.1523/JNEUROSCI.1274-22.2023. Epub 2023 Mar 17.
7
How much of reinforcement learning is working memory, not reinforcement learning? A behavioral, computational, and neurogenetic analysis.强化学习中有多少是工作记忆,而不是强化学习?一项行为、计算和神经遗传学分析。
Eur J Neurosci. 2012 Apr;35(7):1024-35. doi: 10.1111/j.1460-9568.2011.07980.x.
8
Working memory contributions to reinforcement learning impairments in schizophrenia.工作记忆对精神分裂症强化学习障碍的影响
J Neurosci. 2014 Oct 8;34(41):13747-56. doi: 10.1523/JNEUROSCI.0989-14.2014.
9
Relevance of working memory for reinforcement learning in older adults varies with timescale of learning.工作记忆与老年人强化学习的相关性随学习时间尺度而异。
Neuropsychol Dev Cogn B Aging Neuropsychol Cogn. 2020 Sep;27(5):654-676. doi: 10.1080/13825585.2019.1664389. Epub 2019 Sep 22.
10
A new computational account of cognitive control over reinforcement-based decision-making: Modeling of a probabilistic learning task.一种关于对基于强化的决策进行认知控制的新计算解释:概率学习任务的建模
Neural Netw. 2015 Nov;71:112-23. doi: 10.1016/j.neunet.2015.08.006. Epub 2015 Aug 20.

引用本文的文献

1
Anterior Cingulate Cortex Causally Supports Meta-Learning.前扣带回皮层因果性地支持元学习。
bioRxiv. 2024 Jun 13:2024.06.12.598723. doi: 10.1101/2024.06.12.598723.
2
Fluoxetine degrades luminance perceptual thresholds while enhancing motivation and reward sensitivity.氟西汀在提高动机和奖励敏感性的同时,会降低亮度感知阈值。
Front Pharmacol. 2023 Apr 20;14:1103999. doi: 10.3389/fphar.2023.1103999. eCollection 2023.
3
The Role of Executive Function in Shaping Reinforcement Learning.执行功能在塑造强化学习中的作用。
Curr Opin Behav Sci. 2021 Apr;38:66-73. doi: 10.1016/j.cobeha.2020.10.003. Epub 2020 Nov 14.
4
Learning at Variable Attentional Load Requires Cooperation of Working Memory, Meta-learning, and Attention-augmented Reinforcement Learning.在可变注意力负荷下学习需要工作记忆、元学习和注意力增强的强化学习的合作。
J Cogn Neurosci. 2021 Dec 6;34(1):79-107. doi: 10.1162/jocn_a_01780.
5
Beyond dichotomies in reinforcement learning.超越强化学习中的二分法。
Nat Rev Neurosci. 2020 Oct;21(10):576-586. doi: 10.1038/s41583-020-0355-6. Epub 2020 Sep 1.
6
Interactions of spatial strategies producing generalization gradient and blocking: A computational approach.产生泛化梯度和阻断的空间策略相互作用:一种计算方法。
PLoS Comput Biol. 2018 Apr 9;14(4):e1006092. doi: 10.1371/journal.pcbi.1006092. eCollection 2018 Apr.