Suppr超能文献

神经放电率操作性条件反射过程中奖励与无监督学习的共存

Coexistence of reward and unsupervised learning during the operant conditioning of neural firing rates.

作者信息

Kerr Robert R, Grayden David B, Thomas Doreen A, Gilson Matthieu, Burkitt Anthony N

机构信息

NeuroEngineering Laboratory, Department of Electrical and Electronic Engineering, University of Melbourne, Melbourne, Australia ; Centre for Neural Engineering, University of Melbourne, Melbourne, Australia ; NICTA, Victoria Research Lab, University of Melbourne, Melbourne, Australia.

NeuroEngineering Laboratory, Department of Electrical and Electronic Engineering, University of Melbourne, Melbourne, Australia ; Centre for Neural Engineering, University of Melbourne, Melbourne, Australia ; NICTA, Victoria Research Lab, University of Melbourne, Melbourne, Australia ; Bionics Institute, Melbourne, Australia.

出版信息

PLoS One. 2014 Jan 27;9(1):e87123. doi: 10.1371/journal.pone.0087123. eCollection 2014.

Abstract

A fundamental goal of neuroscience is to understand how cognitive processes, such as operant conditioning, are performed by the brain. Typical and well studied examples of operant conditioning, in which the firing rates of individual cortical neurons in monkeys are increased using rewards, provide an opportunity for insight into this. Studies of reward-modulated spike-timing-dependent plasticity (RSTDP), and of other models such as R-max, have reproduced this learning behavior, but they have assumed that no unsupervised learning is present (i.e., no learning occurs without, or independent of, rewards). We show that these models cannot elicit firing rate reinforcement while exhibiting both reward learning and ongoing, stable unsupervised learning. To fix this issue, we propose a new RSTDP model of synaptic plasticity based upon the observed effects that dopamine has on long-term potentiation and depression (LTP and LTD). We show, both analytically and through simulations, that our new model can exhibit unsupervised learning and lead to firing rate reinforcement. This requires that the strengthening of LTP by the reward signal is greater than the strengthening of LTD and that the reinforced neuron exhibits irregular firing. We show the robustness of our findings to spike-timing correlations, to the synaptic weight dependence that is assumed, and to changes in the mean reward. We also consider our model in the differential reinforcement of two nearby neurons. Our model aligns more strongly with experimental studies than previous models and makes testable predictions for future experiments.

摘要

神经科学的一个基本目标是了解大脑如何执行诸如操作性条件反射等认知过程。操作性条件反射的典型且经过充分研究的例子,即通过奖励提高猴子单个皮层神经元的放电率,为深入了解这一过程提供了契机。对奖励调制的尖峰时间依赖性可塑性(RSTDP)以及其他模型(如R-max)的研究重现了这种学习行为,但它们假设不存在无监督学习(即没有奖励或独立于奖励时不发生学习)。我们表明,这些模型在展现奖励学习和持续稳定的无监督学习时,无法引发放电率增强。为解决这个问题,我们基于多巴胺对长时程增强和长时程抑制(LTP和LTD)的观察效应,提出了一种新的突触可塑性RSTDP模型。我们通过分析和模拟表明,我们的新模型能够展现无监督学习并导致放电率增强。这要求奖励信号对LTP的增强大于对LTD的增强,且被增强的神经元表现出不规则放电。我们展示了我们的研究结果对于尖峰时间相关性、假设的突触权重依赖性以及平均奖励变化的稳健性。我们还在两个相邻神经元的差异强化中考虑了我们的模型。与先前的模型相比,我们的模型与实验研究的契合度更高,并为未来的实验做出了可检验的预测。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d512/3903641/e7a9c806b4a3/pone.0087123.g001.jpg

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验