Suppr超能文献

重新思考多巴胺作为一般性预测误差。

Rethinking dopamine as generalized prediction error.

机构信息

Intramural Research Program of the National Institute on Drug Abuse, NIH, Bethesda, MD, USA.

Department of Anatomy and Neurobiology, University of Maryland School of Medicine, Baltimore, MD, USA.

出版信息

Proc Biol Sci. 2018 Nov 21;285(1891):20181645. doi: 10.1098/rspb.2018.1645.

Abstract

Midbrain dopamine neurons are commonly thought to report a reward prediction error (RPE), as hypothesized by reinforcement learning (RL) theory. While this theory has been highly successful, several lines of evidence suggest that dopamine activity also encodes sensory prediction errors unrelated to reward. Here, we develop a new theory of dopamine function that embraces a broader conceptualization of prediction errors. By signalling errors in both sensory and reward predictions, dopamine supports a form of RL that lies between model-based and model-free algorithms. This account remains consistent with current canon regarding the correspondence between dopamine transients and RPEs, while also accounting for new data suggesting a role for these signals in phenomena such as sensory preconditioning and identity unblocking, which ostensibly draw upon knowledge beyond reward predictions.

摘要

中脑多巴胺神经元通常被认为报告了奖励预测误差(RPE),这是强化学习(RL)理论所假设的。虽然该理论非常成功,但有几条证据表明,多巴胺活动也编码了与奖励无关的感觉预测误差。在这里,我们提出了一个新的多巴胺功能理论,它包含了对预测误差的更广泛的概念化。通过对感觉和奖励预测中的误差进行信号传递,多巴胺支持了一种介于基于模型和无模型算法之间的 RL。该解释与当前关于多巴胺瞬变与 RPE 之间对应关系的规范仍然一致,同时也解释了新的数据,表明这些信号在诸如感觉预处理和身份解锁等现象中发挥作用,这些现象显然依赖于超出奖励预测的知识。

相似文献

1
Rethinking dopamine as generalized prediction error.重新思考多巴胺作为一般性预测误差。
Proc Biol Sci. 2018 Nov 21;285(1891):20181645. doi: 10.1098/rspb.2018.1645.
2
A dopamine mechanism for reward maximization.多巴胺奖赏最大化机制。
Proc Natl Acad Sci U S A. 2024 May 14;121(20):e2316658121. doi: 10.1073/pnas.2316658121. Epub 2024 May 8.
6
Belief state representation in the dopamine system.多巴胺系统中的信念状态表示。
Nat Commun. 2018 May 14;9(1):1891. doi: 10.1038/s41467-018-04397-0.

引用本文的文献

5
The interoceptive origin of reinforcement learning.强化学习的内感受起源
Trends Cogn Sci. 2025 Sep;29(9):840-854. doi: 10.1016/j.tics.2025.05.008. Epub 2025 Jun 10.
8
Devaluing memories of reward: a case for dopamine.贬低奖励记忆:多巴胺的一个实例
Commun Biol. 2025 Feb 3;8(1):161. doi: 10.1038/s42003-024-07440-7.

本文引用的文献

2
The successor representation in human reinforcement learning.人类强化学习中的后继表示
Nat Hum Behav. 2017 Sep;1(9):680-692. doi: 10.1038/s41562-017-0180-8. Epub 2017 Aug 28.
4
The Successor Representation: Its Computational Logic and Neural Substrates.后继者表象:其计算逻辑与神经基质。
J Neurosci. 2018 Aug 15;38(33):7193-7200. doi: 10.1523/JNEUROSCI.0151-18.2018. Epub 2018 Jul 13.
7
Model-based predictions for dopamine.基于模型的多巴胺预测。
Curr Opin Neurobiol. 2018 Apr;49:1-7. doi: 10.1016/j.conb.2017.10.006. Epub 2017 Oct 31.
9
The hippocampus as a predictive map.海马体作为一个预测图。
Nat Neurosci. 2017 Nov;20(11):1643-1653. doi: 10.1038/nn.4650. Epub 2017 Oct 2.
10
Dopamine, Inference, and Uncertainty.多巴胺、推理与不确定性。
Neural Comput. 2017 Dec;29(12):3311-3326. doi: 10.1162/neco_a_01023. Epub 2017 Sep 28.

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验