• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

相似文献

1
Reinforcement learning with Marr.与马尔理论相结合的强化学习
Curr Opin Behav Sci. 2016 Oct;11:67-73. doi: 10.1016/j.cobeha.2016.04.005.
2
An assessment of Marr's theory of the hippocampus as a temporary memory store.对马尔关于海马体作为临时记忆存储库的理论的评估。
Philos Trans R Soc Lond B Biol Sci. 1990 Aug 29;329(1253):205-15. doi: 10.1098/rstb.1990.0165.
3
Marr's Levels Revisited: Understanding How Brains Break.重温马尔的层次理论:理解大脑如何崩溃。
Top Cogn Sci. 2015 Apr;7(2):259-73. doi: 10.1111/tops.12130. Epub 2015 Apr 23.
4
Marr's Attacks: On Reductionism and Vagueness.马尔的抨击:论还原论与模糊性
Top Cogn Sci. 2015 Apr;7(2):323-35. doi: 10.1111/tops.12133. Epub 2015 Mar 5.
5
Marr and reductionism.马尔与还原论。
Top Cogn Sci. 2015 Apr;7(2):299-311. doi: 10.1111/tops.12134. Epub 2015 Mar 13.
6
The Non-Redundant Contributions of Marr's Three Levels of Analysis for Explaining Information-Processing Mechanisms.马尔解释信息处理机制的三个分析层次的非冗余贡献
Top Cogn Sci. 2015 Apr;7(2):312-22. doi: 10.1111/tops.12141. Epub 2015 Apr 22.
7
Does this computational theory solve the right problem? Marr, Gibson, and the goal of vision.这种计算理论解决的是正确的问题吗?马尔、吉布森与视觉的目标。
Perception. 2012;41(9):1053-60. doi: 10.1068/p7327.
8
The algorithmic level is the bridge between computation and brain.算法层面是计算与大脑之间的桥梁。
Top Cogn Sci. 2015 Apr;7(2):230-42. doi: 10.1111/tops.12131. Epub 2015 Mar 30.
9
Thirty Years After Marr's Vision: Levels of Analysis in Cognitive Science.马尔的视觉理论提出三十年之后:认知科学中的分析层次
Top Cogn Sci. 2015 Apr;7(2):187-90. doi: 10.1111/tops.12137. Epub 2015 Mar 9.
10
Marr's theory of the neocortex as a self-organizing neural network.马尔将新皮层视为自组织神经网络的理论。
Neural Comput. 1997 May 15;9(4):911-36. doi: 10.1162/neco.1997.9.4.911.

引用本文的文献

1
Goals and the Structure of Experience.目标与经验的结构
ArXiv. 2025 Aug 20:arXiv:2508.15013v1.
2
Humans learn generalizable representations through efficient coding.人类通过高效编码学习可泛化的表征。
Nat Commun. 2025 Apr 29;16(1):3989. doi: 10.1038/s41467-025-58848-6.
3
Schemas, reinforcement learning and the medial prefrontal cortex.图式、强化学习与内侧前额叶皮质
Nat Rev Neurosci. 2025 Mar;26(3):141-157. doi: 10.1038/s41583-024-00893-z. Epub 2025 Jan 7.
4
Reinforcement-Learning-Informed Queries Guide Behavioral Change.强化学习引导的查询指导行为改变。
Clin Psychol Sci. 2024 Nov;12(6):1146-1161. doi: 10.1177/21677026231213368. Epub 2024 Jan 24.
5
Feasibility of dopamine as a vector-valued feedback signal in the basal ganglia.多巴胺作为基底神经节中向量值反馈信号的可行性。
Proc Natl Acad Sci U S A. 2023 Aug 8;120(32):e2221994120. doi: 10.1073/pnas.2221994120. Epub 2023 Aug 1.
6
A neuro-computational social learning framework to facilitate transdiagnostic classification and treatment across psychiatric disorders.一个神经计算的社会学习框架,以促进跨诊断分类和治疗精神障碍。
Neurosci Biobehav Rev. 2023 Jun;149:105181. doi: 10.1016/j.neubiorev.2023.105181. Epub 2023 Apr 14.
7
Understanding cingulotomy's therapeutic effect in OCD through computer models.通过计算机模型理解扣带束切开术在强迫症中的治疗效果。
Front Integr Neurosci. 2023 Jan 10;16:889831. doi: 10.3389/fnint.2022.889831. eCollection 2022.
8
Incorporating social knowledge structures into computational models.将社会知识结构纳入计算模型中。
Nat Commun. 2022 Oct 20;13(1):6205. doi: 10.1038/s41467-022-33418-2.
9
Efficient coding of cognitive variables underlies dopamine response and choice behavior.认知变量的有效编码是多巴胺反应和选择行为的基础。
Nat Neurosci. 2022 Jun;25(6):738-748. doi: 10.1038/s41593-022-01085-7. Epub 2022 Jun 6.
10
Learning from other minds: An optimistic critique of reinforcement learning models of social learning.向他人学习:对社会学习强化学习模型的乐观批判。
Curr Opin Behav Sci. 2021 Apr;38:110-115. doi: 10.1016/j.cobeha.2021.01.006. Epub 2021 Mar 23.

本文引用的文献

1
Temporal Specificity of Reward Prediction Errors Signaled by Putative Dopamine Neurons in Rat VTA Depends on Ventral Striatum.大鼠腹侧被盖区中假定多巴胺能神经元发出的奖励预测误差的时间特异性取决于腹侧纹状体。
Neuron. 2016 Jul 6;91(1):182-93. doi: 10.1016/j.neuron.2016.05.015. Epub 2016 Jun 9.
2
Characterizing a psychiatric symptom dimension related to deficits in goal-directed control.表征与目标导向控制缺陷相关的一种精神症状维度。
Elife. 2016 Mar 1;5:e11305. doi: 10.7554/eLife.11305.
3
When good news leads to bad choices.当好消息导致错误选择时。
J Exp Anal Behav. 2016 Jan;105(1):23-40. doi: 10.1002/jeab.192.
4
When good pigeons make bad decisions: Choice with probabilistic delays and outcomes.当优秀的鸽子做出错误决策时:具有概率性延迟和结果的选择。
J Exp Anal Behav. 2015 Nov;104(3):241-51. doi: 10.1002/jeab.177.
5
Building health behavior models to guide the development of just-in-time adaptive interventions: A pragmatic framework.构建健康行为模型以指导即时自适应干预措施的开发:一个实用框架。
Health Psychol. 2015 Dec;34S(0):1209-19. doi: 10.1037/hea0000306.
6
Brief optogenetic inhibition of dopamine neurons mimics endogenous negative reward prediction errors.对多巴胺神经元进行短暂的光遗传学抑制可模拟内源性负性奖励预测误差。
Nat Neurosci. 2016 Jan;19(1):111-6. doi: 10.1038/nn.4191. Epub 2015 Dec 7.
7
Action initiation shapes mesolimbic dopamine encoding of future rewards.行动发起塑造了中脑边缘多巴胺对未来奖励的编码。
Nat Neurosci. 2016 Jan;19(1):34-6. doi: 10.1038/nn.4187. Epub 2015 Dec 7.
8
Striatal dynamics explain duration judgments.纹状体动力学解释了时长判断。
Elife. 2015 Dec 7;4:e11386. doi: 10.7554/eLife.11386.
9
Arithmetic and local circuitry underlying dopamine prediction errors.多巴胺预测误差背后的算术和局部神经回路。
Nature. 2015 Sep 10;525(7568):243-6. doi: 10.1038/nature14855. Epub 2015 Aug 31.
10
Reinforcement learning in multidimensional environments relies on attention mechanisms.多维环境中的强化学习依赖于注意力机制。
J Neurosci. 2015 May 27;35(21):8145-57. doi: 10.1523/JNEUROSCI.2978-14.2015.

与马尔理论相结合的强化学习

Reinforcement learning with Marr.

作者信息

Niv Yael, Langdon Angela

机构信息

Psychology Department & Princeton Neuroscience Institute, Princeton University, Princeton, New Jersey, 08540.

出版信息

Curr Opin Behav Sci. 2016 Oct;11:67-73. doi: 10.1016/j.cobeha.2016.04.005.

DOI:10.1016/j.cobeha.2016.04.005
PMID:27408906
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC4939081/
Abstract

To many, the poster child for David Marr's famous three levels of scientific inquiry is reinforcement learning-a computational theory of reward optimization, which readily prescribes algorithmic solutions that evidence striking resemblance to signals found in the brain, suggesting a straightforward neural implementation. Here we review questions that remain open at each level of analysis, concluding that the path forward to their resolution calls for inspiration across levels, rather than a focus on mutual constraints.

摘要

对许多人来说,大卫·马尔著名的三个科学探究层次的典型代表是强化学习——一种奖励优化的计算理论,它很容易给出算法解决方案,这些方案与大脑中发现的信号惊人地相似,这表明有一种直接的神经实现方式。在这里,我们回顾了在每个分析层次上仍然悬而未决的问题,得出结论:解决这些问题的前进道路需要跨层次的启发,而不是专注于相互约束。