• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

奖励关联不能解释猴子的传递性推理表现。

Reward associations do not explain transitive inference performance in monkeys.

机构信息

Department of Psychology, Columbia University, New York, NY 10027, USA.

Zuckerman Mind Brain Behavior Institute, Columbia University, New York, NY 10027, USA.

出版信息

Sci Adv. 2019 Jul 31;5(7):eaaw2089. doi: 10.1126/sciadv.aaw2089. eCollection 2019 Jul.

DOI:10.1126/sciadv.aaw2089
PMID:32128384
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7032924/
Abstract

Most accounts of behavior in nonhuman animals assume that they make choices to maximize expected reward value. However, model-free reinforcement learning based on reward associations cannot account for choice behavior in transitive inference paradigms. We manipulated the amount of reward associated with each item of an ordered list, so that maximizing expected reward value was always in conflict with decision rules based on the implicit list order. Under such a schedule, model-free reinforcement algorithms cannot achieve high levels of accuracy, even after extensive training. Monkeys nevertheless learned to make correct rule-based choices. These results show that monkeys' performance in transitive inference paradigms is not driven solely by expected reward and that appropriate inferences are made despite discordant reward incentives. We show that their choices can be explained by an abstract, model-based representation of list order, and we provide a method for inferring the contents of such representations from observed data.

摘要

大多数关于非人类动物行为的解释都假设它们会做出选择以最大化预期奖励值。然而,基于奖励关联的无模型强化学习无法解释在传递推理范式中的选择行为。我们操纵了与有序列表中每个项目相关联的奖励数量,因此最大化预期奖励值总是与基于隐式列表顺序的决策规则相冲突。在这种情况下,即使经过广泛的训练,无模型强化算法也无法达到很高的准确性。然而,猴子学会了做出正确的基于规则的选择。这些结果表明,猴子在传递推理范式中的表现不仅仅取决于预期奖励,并且尽管奖励激励存在不一致,但仍能做出适当的推断。我们表明,他们的选择可以通过对列表顺序的抽象、基于模型的表示来解释,并且我们提供了一种从观察数据中推断这种表示内容的方法。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8e3c/7032924/40794ca1a3e5/aaw2089-F6.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8e3c/7032924/adb7628de978/aaw2089-F1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8e3c/7032924/250feae93187/aaw2089-F2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8e3c/7032924/61857c96ce09/aaw2089-F3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8e3c/7032924/e22743e4d269/aaw2089-F4.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8e3c/7032924/2728d95b2185/aaw2089-F5.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8e3c/7032924/40794ca1a3e5/aaw2089-F6.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8e3c/7032924/adb7628de978/aaw2089-F1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8e3c/7032924/250feae93187/aaw2089-F2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8e3c/7032924/61857c96ce09/aaw2089-F3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8e3c/7032924/e22743e4d269/aaw2089-F4.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8e3c/7032924/2728d95b2185/aaw2089-F5.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8e3c/7032924/40794ca1a3e5/aaw2089-F6.jpg

相似文献

1
Reward associations do not explain transitive inference performance in monkeys.奖励关联不能解释猴子的传递性推理表现。
Sci Adv. 2019 Jul 31;5(7):eaaw2089. doi: 10.1126/sciadv.aaw2089. eCollection 2019 Jul.
2
Influence of Rule- and Reward-based Strategies on Inferences of Serial Order by Monkeys.基于规则和奖励策略对猴子进行序列推理的影响。
J Cogn Neurosci. 2022 Mar 5;34(4):592-604. doi: 10.1162/jocn_a_01823.
3
Self-choice enhances value in reward-seeking in primates.自我选择增强了灵长类动物在寻求奖励中的价值。
Neurosci Res. 2014 Mar;80:45-54. doi: 10.1016/j.neures.2014.01.004. Epub 2014 Jan 22.
4
Cognitive mechanisms for transitive inference performance in rhesus monkeys: measuring the influence of associative strength and inferred order.恒河猴传递性推理表现的认知机制:测量联想强度和推理顺序的影响
J Exp Psychol Anim Behav Process. 2012 Oct;38(4):331-45. doi: 10.1037/a0030306.
5
Nutrient-Sensitive Reinforcement Learning in Monkeys.猴子的营养敏感强化学习。
J Neurosci. 2023 Mar 8;43(10):1714-1730. doi: 10.1523/JNEUROSCI.0752-22.2022. Epub 2023 Jan 20.
6
Associative models fail to characterize transitive inference performance in rhesus monkeys (Macaca mulatta).联想模型无法描述恒河猴(猕猴)的传递性推理表现。
Learn Behav. 2020 Mar;48(1):135-148. doi: 10.3758/s13420-020-00417-6.
7
Utility functions predict variance and skewness risk preferences in monkeys.效用函数可预测猴子的方差和偏度风险偏好。
Proc Natl Acad Sci U S A. 2016 Jul 26;113(30):8402-7. doi: 10.1073/pnas.1602217113. Epub 2016 Jul 11.
8
Worth the Work? Monkeys Discount Rewards by a Subjective Adapting Effort Cost.值得付出努力吗?猴子通过主观适应努力成本来折扣奖励。
J Neurosci. 2023 Oct 4;43(40):6796-6806. doi: 10.1523/JNEUROSCI.0115-23.2023. Epub 2023 Aug 25.
9
Mechanisms of inferential order judgments in humans (Homo sapiens) and rhesus monkeys (Macaca mulatta).人类(智人)和恒河猴(猕猴)推理顺序判断的机制。
J Comp Psychol. 2011 May;125(2):227-38. doi: 10.1037/a0021572.
10
Positional inference in rhesus macaques.恒河猴中的位置推断。
Anim Cogn. 2022 Feb;25(1):73-93. doi: 10.1007/s10071-021-01536-x. Epub 2021 Jul 24.

引用本文的文献

1
Non-human primates can flexibly learn serial sequences and reorder context-dependent object sequences.非人灵长类动物能够灵活地学习序列,并对依赖上下文的物体序列进行重新排序。
PLoS Biol. 2025 Jun 23;23(6):e3003255. doi: 10.1371/journal.pbio.3003255. eCollection 2025 Jun.
2
Flexible Learning and Re-ordering of Context-dependent Object Sequences in Nonhuman Primates.非人灵长类动物中上下文相关物体序列的灵活学习与重新排序
bioRxiv. 2024 Nov 24:2024.11.24.625056. doi: 10.1101/2024.11.24.625056.
3
A geometrical solution underlies general neural principle for serial ordering.

本文引用的文献

1
Stan: A Probabilistic Programming Language.斯坦:一种概率编程语言。
J Stat Softw. 2017;76. doi: 10.18637/jss.v076.i01. Epub 2017 Jan 11.
2
Inferential Learning of Serial Order of Perceptual Categories by Rhesus Monkeys ().恒河猴对感知类别序列顺序的推理学习()
J Neurosci. 2017 Jun 28;37(26):6268-6276. doi: 10.1523/JNEUROSCI.0263-17.2017. Epub 2017 May 25.
3
Transitive inference in humans (Homo sapiens) and rhesus macaques (Macaca mulatta) after massed training of the last two list items.在对最后两个列表项进行集中训练后,人类(智人)和恒河猴(猕猴)的传递性推理。
一种几何解为序列排序的一般神经原则提供了基础。
Nat Commun. 2024 Sep 19;15(1):8238. doi: 10.1038/s41467-024-52240-6.
4
Transitive inference in a clinical childhood sample with a focus on autism spectrum disorder.临床儿童样本中转译推理研究,重点关注自闭症谱系障碍。
Autism Res. 2024 Nov;17(11):2355-2369. doi: 10.1002/aur.3225. Epub 2024 Sep 2.
5
Probabilistic reinforcement precludes transitive inference: A preliminary study.概率强化排除传递性推理:一项初步研究。
Front Psychol. 2023 Mar 30;14:1111597. doi: 10.3389/fpsyg.2023.1111597. eCollection 2023.
6
Prior experience modifies acquisition trajectories via response-strategy sampling.先前经验通过反应策略抽样来改变习得轨迹。
Anim Cogn. 2023 Jul;26(4):1217-1239. doi: 10.1007/s10071-023-01769-y. Epub 2023 Apr 10.
7
Resolving the associative learning paradox by category learning in pigeons.通过鸽子的类别学习解决联想学习悖论。
Curr Biol. 2023 Mar 27;33(6):1112-1116.e2. doi: 10.1016/j.cub.2023.01.024. Epub 2023 Feb 7.
8
Hippocampal and medial prefrontal cortices encode structural task representations following progressive and interleaved training schedules.海马体和内侧前额叶皮质在渐进式和交错式训练计划后,对结构任务的表现进行编码。
PLoS Comput Biol. 2022 Oct 17;18(10):e1010566. doi: 10.1371/journal.pcbi.1010566. eCollection 2022 Oct.
9
Superstitious learning of abstract order from random reinforcement.从随机强化中迷信地学习抽象秩序。
Proc Natl Acad Sci U S A. 2022 Aug 30;119(35):e2202789119. doi: 10.1073/pnas.2202789119. Epub 2022 Aug 23.
10
Flexible auditory training, psychophysics, and enrichment of common marmosets with an automated, touchscreen-based system.采用自动化、基于触摸屏的系统对普通狨猴进行灵活的听觉训练、心理物理学和丰富化。
Nat Commun. 2022 Mar 28;13(1):1648. doi: 10.1038/s41467-022-29185-9.
J Comp Psychol. 2017 Aug;131(3):231-245. doi: 10.1037/com0000065. Epub 2017 Mar 23.
4
Implicit Value Updating Explains Transitive Inference Performance: The Betasort Model.内隐价值更新解释传递性推理表现:贝塔排序模型。
PLoS Comput Biol. 2015 Sep 25;11(9):e1004523. doi: 10.1371/journal.pcbi.1004523. eCollection 2015.
5
Effects of spatial training on transitive inference performance in humans and rhesus monkeys.空间训练对人类和恒河猴传递性推理表现的影响。
J Exp Psychol Anim Learn Cogn. 2014 Oct;40(4):477-89. doi: 10.1037/xan0000038. Epub 2014 Jul 28.
6
Assessing value representation in animals.评估动物的价值表征。
J Physiol Paris. 2015 Feb-Jun;109(1-3):64-9. doi: 10.1016/j.jphysparis.2014.07.003. Epub 2014 Aug 1.
7
Transfer of a serial representation between two distinct tasks by rhesus macaques.恒河猴在两个不同任务之间的串行表示的转移。
PLoS One. 2013 Jul 31;8(7):e70285. doi: 10.1371/journal.pone.0070285. Print 2013.
8
Orbitofrontal cortex supports behavior and learning using inferred but not cached values.眶额皮层使用推断值而不是缓存值来支持行为和学习。
Science. 2012 Nov 16;338(6109):953-6. doi: 10.1126/science.1227489.
9
Transitive inference in pigeons: measuring the associative values of Stimuli B and D.鸽子的传递性推理:测量刺激B和D的联想值。
Behav Processes. 2012 Mar;89(3):244-55. doi: 10.1016/j.beproc.2011.12.001. Epub 2011 Dec 15.
10
Multiple representations and algorithms for reinforcement learning in the cortico-basal ganglia circuit.皮质基底神经节回路中的强化学习的多种表示和算法。
Curr Opin Neurobiol. 2011 Jun;21(3):368-73. doi: 10.1016/j.conb.2011.04.001. Epub 2011 Apr 29.