• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

多巴胺能神经元可以表征依赖于情境的预测误差。

Dopamine neurons can represent context-dependent prediction error.

作者信息

Nakahara Hiroyuki, Itoh Hideaki, Kawagoe Reiko, Takikawa Yoriko, Hikosaka Okihide

机构信息

Lab for Mathematical Neuroscience, RIKEN Brain Science Institute, Wako, Saitama, Japan.

出版信息

Neuron. 2004 Jan 22;41(2):269-80. doi: 10.1016/s0896-6273(03)00869-9.

DOI:10.1016/s0896-6273(03)00869-9
PMID:14741107
Abstract

Midbrain dopamine (DA) neurons are thought to encode reward prediction error. Reward prediction can be improved if any relevant context is taken into account. We found that monkey DA neurons can encode a context-dependent prediction error. In the first noncontextual task, a light stimulus was randomly followed by reward, with a fixed equal probability. The response of DA neurons was positively correlated with the number of preceding unrewarded trials and could be simulated by a conventional temporal difference (TD) model. In the second contextual task, a reward-indicating light stimulus was presented with the probability that, while fixed overall, was incremented as a function of the number of preceding unrewarded trials. The DA neuronal response then was negatively correlated with this number. This history effect corresponded to the prediction error based on the conditional probability of reward and could be simulated only by implementing the relevant context into the TD model.

摘要

中脑多巴胺(DA)神经元被认为编码奖励预测误差。如果考虑任何相关背景,奖励预测可以得到改善。我们发现猴子的DA神经元可以编码依赖于背景的预测误差。在第一个非背景任务中,光刺激之后随机给予奖励,概率固定且相等。DA神经元的反应与之前未获得奖励的试验次数呈正相关,并且可以用传统的时间差分(TD)模型进行模拟。在第二个背景任务中,呈现一个指示奖励的光刺激,其概率虽然总体上是固定的,但会根据之前未获得奖励的试验次数而增加。然后DA神经元的反应与这个次数呈负相关。这种历史效应对应于基于奖励条件概率的预测误差,并且只有在将相关背景纳入TD模型时才能进行模拟。

相似文献

1
Dopamine neurons can represent context-dependent prediction error.多巴胺能神经元可以表征依赖于情境的预测误差。
Neuron. 2004 Jan 22;41(2):269-80. doi: 10.1016/s0896-6273(03)00869-9.
2
Midbrain dopamine neurons encode a quantitative reward prediction error signal.中脑多巴胺神经元编码一种定量奖励预测误差信号。
Neuron. 2005 Jul 7;47(1):129-41. doi: 10.1016/j.neuron.2005.05.020.
3
Responses of monkey dopamine neurons during learning of behavioral reactions.猴子多巴胺神经元在行为反应学习过程中的反应
J Neurophysiol. 1992 Jan;67(1):145-63. doi: 10.1152/jn.1992.67.1.145.
4
Axiomatic methods, dopamine and reward prediction error.公理法、多巴胺与奖励预测误差。
Curr Opin Neurobiol. 2008 Apr;18(2):197-202. doi: 10.1016/j.conb.2008.07.007. Epub 2008 Aug 12.
5
Midbrain dopamine neurons compute inferred and cached value prediction errors in a common framework.中脑多巴胺神经元在一个通用框架中计算推断和缓存的价值预测误差。
Elife. 2016 Mar 7;5:e13665. doi: 10.7554/eLife.13665.
6
Midbrain dopaminergic neurons and striatal cholinergic interneurons encode the difference between reward and aversive events at different epochs of probabilistic classical conditioning trials.中脑多巴胺能神经元和纹状体胆碱能中间神经元在概率性经典条件反射试验的不同阶段编码奖励与厌恶事件之间的差异。
J Neurosci. 2008 Nov 5;28(45):11673-84. doi: 10.1523/JNEUROSCI.3839-08.2008.
7
A possible role of midbrain dopamine neurons in short- and long-term adaptation of saccades to position-reward mapping.中脑多巴胺能神经元在扫视运动对位置-奖励映射的短期和长期适应中的可能作用。
J Neurophysiol. 2004 Oct;92(4):2520-9. doi: 10.1152/jn.00238.2004. Epub 2004 May 26.
8
Responses of monkey midbrain dopamine neurons during delayed alternation performance.猴子在延迟交替任务执行过程中中脑多巴胺能神经元的反应。
Brain Res. 1991 Dec 20;567(2):337-41. doi: 10.1016/0006-8993(91)90816-e.
9
Adolescent Dopamine Neurons Represent Reward Differently during Action and State Guided Learning.青少年多巴胺神经元在动作和状态引导学习中对奖励的表现不同。
J Neurosci. 2021 Nov 10;41(45):9419-9430. doi: 10.1523/JNEUROSCI.1321-21.2021. Epub 2021 Oct 5.
10
Correlated coding of motivation and outcome of decision by dopamine neurons.多巴胺神经元对动机与决策结果的关联编码。
J Neurosci. 2003 Oct 29;23(30):9913-23. doi: 10.1523/JNEUROSCI.23-30-09913.2003.

引用本文的文献

1
Dopamine neurons encode trial-by-trial subjective reward value in an auction-like task.在类似拍卖的任务中,多巴胺神经元对每次试验的主观奖励值进行编码。
Nat Commun. 2024 Sep 17;15(1):8138. doi: 10.1038/s41467-024-52311-8.
2
Temporal regularities shape perceptual decisions and striatal dopamine signals.时间规律塑造知觉决策和纹状体多巴胺信号。
Nat Commun. 2024 Aug 17;15(1):7093. doi: 10.1038/s41467-024-51393-8.
3
A Basal Ganglia model for understanding working memory functions in healthy and Parkinson's conditions.一个用于理解健康和帕金森病状态下工作记忆功能的基底神经节模型。
Cogn Neurodyn. 2024 Aug;18(4):1913-1929. doi: 10.1007/s11571-023-10056-y. Epub 2024 Jan 9.
4
Dopamine-independent effect of rewards on choices through hidden-state inference.奖励对选择的多巴胺非依赖效应:通过隐状态推断。
Nat Neurosci. 2024 Feb;27(2):286-297. doi: 10.1038/s41593-023-01542-x. Epub 2024 Jan 12.
5
Dual credit assignment processes underlie dopamine signals in a complex spatial environment.双信用分配过程是复杂空间环境中多巴胺信号的基础。
Neuron. 2023 Nov 1;111(21):3465-3478.e7. doi: 10.1016/j.neuron.2023.07.017. Epub 2023 Aug 22.
6
A stable, distributed code for cue value in mouse cortex during reward learning.在奖励学习过程中,小鼠皮层中线索值的稳定、分布式编码。
Elife. 2023 Jun 30;12:RP84604. doi: 10.7554/eLife.84604.
7
What is dopamine doing in model-based reinforcement learning?多巴胺在基于模型的强化学习中起什么作用?
Curr Opin Behav Sci. 2021 Apr;38:74-82. doi: 10.1016/j.cobeha.2020.10.010.
8
Dual credit assignment processes underlie dopamine signals in a complex spatial environment.在复杂空间环境中,双重学分分配过程是多巴胺信号的基础。
bioRxiv. 2023 Mar 19:2023.02.15.528738. doi: 10.1101/2023.02.15.528738.
9
Dopamine neurons encode trial-by-trial subjective reward value in an auction-like task.多巴胺能神经元在类似拍卖的任务中逐次编码主观奖励值。
bioRxiv. 2024 May 10:2023.01.20.524896. doi: 10.1101/2023.01.20.524896.
10
Importance of prefrontal meta control in human-like reinforcement learning.前额叶元控制在类人强化学习中的重要性。
Front Comput Neurosci. 2022 Dec 21;16:1060101. doi: 10.3389/fncom.2022.1060101. eCollection 2022.