• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

人类重复使用策略来解决新任务的神经学证据。

Neural evidence that humans reuse strategies to solve new tasks.

作者信息

Hall-McMaster Sam, Tomov Momchil S, Gershman Samuel J, Schuck Nicolas W

机构信息

Department of Psychology and Center for Brain Science, Harvard University, Cambridge, Massachusetts, United States of America.

Max Planck Institute for Human Development, Berlin, Germany.

出版信息

PLoS Biol. 2025 Jun 5;23(6):e3003174. doi: 10.1371/journal.pbio.3003174. eCollection 2025 Jun.

DOI:10.1371/journal.pbio.3003174
PMID:40471906
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12140246/
Abstract

Generalization from past experience is an important feature of intelligent systems. When faced with a new task, one efficient computational approach is to evaluate solutions to earlier tasks as candidates for reuse. Consistent with this idea, we found that human participants (n = 38) learned optimal solutions to a set of training tasks and generalized them to novel test tasks in a reward-selective manner. This behavior was consistent with a computational process based on the successor representation known as successor features and generalized policy improvement (SF&GPI). Neither model-free perseveration or model-based control using a complete model of the environment could explain choice behavior. Decoding from functional magnetic resonance imaging data revealed that solutions from the SF&GPI algorithm were activated on test tasks in visual and prefrontal cortex. This activation had a functional connection to behavior in that stronger activation of SF&GPI solutions in visual areas was associated with increased behavioral reuse. These findings point to a possible neural implementation of an adaptive algorithm for generalization across tasks.

摘要

从过去的经验中进行归纳是智能系统的一个重要特征。当面对一项新任务时,一种有效的计算方法是将早期任务的解决方案作为可复用的候选方案进行评估。与这一观点一致,我们发现人类参与者(n = 38)学习了一组训练任务的最优解决方案,并以奖励选择性的方式将其推广到新的测试任务中。这种行为与基于后继表示的计算过程一致,该过程被称为后继特征和广义策略改进(SF&GPI)。无模型的固执或使用环境完整模型的基于模型的控制都无法解释选择行为。从功能磁共振成像数据进行解码显示,SF&GPI算法的解决方案在测试任务中于视觉和前额叶皮层被激活。这种激活与行为存在功能联系,即视觉区域中SF&GPI解决方案的更强激活与行为复用的增加相关。这些发现指出了一种跨任务归纳的自适应算法可能的神经实现方式。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cbf6/12140246/d7a23d27f959/pbio.3003174.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cbf6/12140246/fe46e9f6ed02/pbio.3003174.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cbf6/12140246/25a336445fe2/pbio.3003174.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cbf6/12140246/ad06bb9ff80e/pbio.3003174.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cbf6/12140246/d7a23d27f959/pbio.3003174.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cbf6/12140246/fe46e9f6ed02/pbio.3003174.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cbf6/12140246/25a336445fe2/pbio.3003174.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cbf6/12140246/ad06bb9ff80e/pbio.3003174.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cbf6/12140246/d7a23d27f959/pbio.3003174.g004.jpg

相似文献

1
Neural evidence that humans reuse strategies to solve new tasks.人类重复使用策略来解决新任务的神经学证据。
PLoS Biol. 2025 Jun 5;23(6):e3003174. doi: 10.1371/journal.pbio.3003174. eCollection 2025 Jun.
2
Reward-dependent learning in neuronal networks for planning and decision making.用于规划和决策的神经网络中基于奖励的学习。
Prog Brain Res. 2000;126:217-29. doi: 10.1016/S0079-6123(00)26016-0.
3
No Effect of Commercial Cognitive Training on Brain Activity, Choice Behavior, or Cognitive Performance.商业认知训练对大脑活动、选择行为或认知表现无影响。
J Neurosci. 2017 Aug 2;37(31):7390-7402. doi: 10.1523/JNEUROSCI.2832-16.2017. Epub 2017 Jul 10.
4
Tonic or Phasic Stimulation of Dopaminergic Projections to Prefrontal Cortex Causes Mice to Maintain or Deviate from Previously Learned Behavioral Strategies.对前额叶皮层多巴胺能投射的强直或相位刺激使小鼠维持或偏离先前习得的行为策略。
J Neurosci. 2017 Aug 30;37(35):8315-8329. doi: 10.1523/JNEUROSCI.1221-17.2017. Epub 2017 Jul 24.
5
Ventrolateral prefrontal cortex is required for performance of a strategy implementation task but not reinforcer devaluation effects in rhesus monkeys.腹外侧前额叶皮层是恒河猴执行策略实施任务所必需的,但对强化物贬值效应并非如此。
Eur J Neurosci. 2009 May;29(10):2049-59. doi: 10.1111/j.1460-9568.2009.06740.x. Epub 2009 May 9.
6
Reward-Guided Learning with and without Causal Attribution.有无因果归因的奖励引导学习
Neuron. 2016 Apr 6;90(1):177-90. doi: 10.1016/j.neuron.2016.02.018. Epub 2016 Mar 10.
7
Rule activation and ventromedial prefrontal engagement support accurate stopping in self-paced learning.规则激活和腹内侧前额叶的参与支持自我调节学习中的准确停止。
Neuroimage. 2018 May 15;172:415-426. doi: 10.1016/j.neuroimage.2018.01.084. Epub 2018 Feb 2.
8
The Neural Representation of Voluntary Task-Set Selection in Dynamic Environments.动态环境中自愿任务集选择的神经表征
Cereb Cortex. 2015 Dec;25(12):4715-26. doi: 10.1093/cercor/bhu155. Epub 2014 Jul 17.
9
Striatal-frontal network activation during voluntary task selection under conditions of monetary reward.纹状体-额皮质网络在货币奖励条件下的自愿任务选择中的激活。
Cogn Affect Behav Neurosci. 2019 Jun;19(3):568-585. doi: 10.3758/s13415-019-00689-0.
10
Reward-related reversal learning after surgical excisions in orbito-frontal or dorsolateral prefrontal cortex in humans.人类眶额皮质或背外侧前额叶皮质手术切除后的奖赏相关反转学习。
J Cogn Neurosci. 2004 Apr;16(3):463-78. doi: 10.1162/089892904322926791.

本文引用的文献

1
Decomposing dynamical subprocesses for compositional generalization.分解动态子过程以实现组合泛化。
Proc Natl Acad Sci U S A. 2024 Nov 12;121(46):e2408134121. doi: 10.1073/pnas.2408134121. Epub 2024 Nov 8.
2
Predictive Representations: Building Blocks of Intelligence.预测表示:智能的基石。
Neural Comput. 2024 Oct 11;36(11):2225-2298. doi: 10.1162/neco_a_01705.
3
Goal-seeking compresses neural codes for space in the human hippocampus and orbitofrontal cortex.目标导向压缩了人类海马体和眶额皮质中空间的神经编码。
Neuron. 2023 Dec 6;111(23):3885-3899.e6. doi: 10.1016/j.neuron.2023.08.021. Epub 2023 Sep 18.
4
TemplateFlow: FAIR-sharing of multi-scale, multi-species brain models.模板流:多尺度、多物种大脑模型的 FAIR 共享。
Nat Methods. 2022 Dec;19(12):1568-1571. doi: 10.1038/s41592-022-01681-2. Epub 2022 Dec 1.
5
Curriculum learning for human compositional generalization.人类组合泛化的课程学习。
Proc Natl Acad Sci U S A. 2022 Oct 11;119(41):e2205582119. doi: 10.1073/pnas.2205582119. Epub 2022 Oct 3.
6
Taking stock of value in the orbitofrontal cortex.审视眶额皮质中的价值。
Nat Rev Neurosci. 2022 Jul;23(7):428-438. doi: 10.1038/s41583-022-00589-2. Epub 2022 Apr 25.
7
Orthogonal representations for robust context-dependent task performance in brains and neural networks.大脑和神经网络中鲁棒上下文相关任务性能的正交表示。
Neuron. 2022 Apr 6;110(7):1258-1270.e11. doi: 10.1016/j.neuron.2022.01.005. Epub 2022 Jan 31.
8
The whole prefrontal cortex is premotor cortex.整个前额皮质都是运动前皮质。
Philos Trans R Soc Lond B Biol Sci. 2022 Feb 14;377(1844):20200524. doi: 10.1098/rstb.2020.0524. Epub 2021 Dec 27.
9
Generating Options and Choosing Between Them Depend on Distinct Forms of Value Representation.生成选项并在它们之间进行选择取决于不同形式的价值表征。
Psychol Sci. 2021 Nov;32(11):1731-1746. doi: 10.1177/09567976211005702. Epub 2021 Sep 27.
10
Temporal and state abstractions for efficient learning, transfer, and composition in humans.人类高效学习、迁移和组合的时间和状态抽象。
Psychol Rev. 2021 Jul;128(4):643-666. doi: 10.1037/rev0000295. Epub 2021 May 20.