• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

在概率任务中,选择之间的时间流逝与重复相同的决策相关。

Time elapsed between choices in a probabilistic task correlates with repeating the same decision.

机构信息

Department of Molecular Neuropharmacology, Maj Institute of Pharmacology, Polish Academy of Sciences, Krakow, Poland.

Department of Structure Research of Condensed Matter, The Henryk Niewodniczański Institute of Nuclear Physics, Polish Academy of Sciences, Krakow, Poland.

出版信息

Eur J Neurosci. 2021 Apr;53(8):2639-2654. doi: 10.1111/ejn.15144. Epub 2021 Mar 2.

DOI:10.1111/ejn.15144
PMID:33559232
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8248175/
Abstract

Reinforcement learning causes an action that yields a positive outcome more likely to be taken in the future. Here, we investigate how the time elapsed from an action affects subsequent decisions. Groups of C57BL6/J mice were housed in IntelliCages with access to water and chow ad libitum; they also had access to bottles with a reward: saccharin solution, alcohol, or a mixture of the two. The probability of receiving a reward in two of the cage corners changed between 0.9 and 0.3 every 48 hr over a period of ~33 days. As expected, in most animals, the odds of repeating a corner choice were increased if that choice was previously rewarded. Interestingly, the time elapsed from the previous choice also influenced the probability of repeating the choice, and this effect was independent of previous outcome. Behavioral data were fitted to a series of reinforcement learning models. Best fits were achieved when the reward prediction update was coupled with separate learning rates from positive and negative outcomes and additionally a "fictitious" update of the expected value of the nonselected choice. Additional inclusion of a time-dependent decay of the expected values improved the fit marginally in some cases.

摘要

强化学习会导致未来更有可能采取产生积极结果的行动。在这里,我们研究了从一个动作到下一个动作的时间间隔是如何影响后续决策的。将 C57BL6/J 小鼠分组放入智能笼中,它们可以自由接触水和食物;它们还可以接触装有奖励的瓶子:糖精溶液、酒精或两者的混合物。在大约 33 天的时间里,两个笼子角落的奖励概率每 48 小时从 0.9 变为 0.3。正如预期的那样,在大多数动物中,如果之前的选择得到了奖励,那么重复选择角落的可能性就会增加。有趣的是,从上次选择到再次选择的时间间隔也会影响选择的概率,而这种影响与之前的结果无关。将行为数据拟合到一系列强化学习模型中。当奖励预测更新与来自正、负结果的单独学习率以及非选择选择的预期价值的“虚拟”更新相结合时,得到了最佳拟合。在某些情况下,额外包含预期值的时间相关衰减略微改善了拟合度。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d51/8248175/cc1e989a2593/EJN-53-2639-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d51/8248175/142b51bba7d9/EJN-53-2639-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d51/8248175/4b077adc0fb1/EJN-53-2639-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d51/8248175/3765dbb95bb9/EJN-53-2639-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d51/8248175/afe2b7af2bb0/EJN-53-2639-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d51/8248175/cdb358745075/EJN-53-2639-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d51/8248175/a4257ce7fb15/EJN-53-2639-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d51/8248175/cc1e989a2593/EJN-53-2639-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d51/8248175/142b51bba7d9/EJN-53-2639-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d51/8248175/4b077adc0fb1/EJN-53-2639-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d51/8248175/3765dbb95bb9/EJN-53-2639-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d51/8248175/afe2b7af2bb0/EJN-53-2639-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d51/8248175/cdb358745075/EJN-53-2639-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d51/8248175/a4257ce7fb15/EJN-53-2639-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d51/8248175/cc1e989a2593/EJN-53-2639-g004.jpg

相似文献

1
Time elapsed between choices in a probabilistic task correlates with repeating the same decision.在概率任务中,选择之间的时间流逝与重复相同的决策相关。
Eur J Neurosci. 2021 Apr;53(8):2639-2654. doi: 10.1111/ejn.15144. Epub 2021 Mar 2.
2
Value-Based Choice, Contingency Learning, and Suicidal Behavior in Mid- and Late-Life Depression.中老年人抑郁症中的基于价值的选择、权变学习与自杀行为。
Biol Psychiatry. 2019 Mar 15;85(6):506-516. doi: 10.1016/j.biopsych.2018.10.006. Epub 2018 Oct 18.
3
Impaired Expected Value Computations Coupled With Overreliance on Stimulus-Response Learning in Schizophrenia.精神分裂症患者的预期价值计算受损,同时过度依赖刺激-反应学习。
Biol Psychiatry Cogn Neurosci Neuroimaging. 2018 Nov;3(11):916-926. doi: 10.1016/j.bpsc.2018.03.014. Epub 2018 Apr 3.
4
Credit Assignment in a Motor Decision Making Task Is Influenced by Agency and Not Sensory Prediction Errors.在一项运动决策任务中,信用分配受机构影响,而不受感官预测误差影响。
J Neurosci. 2018 May 9;38(19):4521-4530. doi: 10.1523/JNEUROSCI.3601-17.2018. Epub 2018 Apr 12.
5
Negative symptoms and the failure to represent the expected reward value of actions: behavioral and computational modeling evidence.阴性症状与无法表征预期的行动奖励价值:行为及计算建模证据
Arch Gen Psychiatry. 2012 Feb;69(2):129-38. doi: 10.1001/archgenpsychiatry.2011.1269.
6
The ubiquity of model-based reinforcement learning.基于模型的强化学习无处不在。
Curr Opin Neurobiol. 2012 Dec;22(6):1075-81. doi: 10.1016/j.conb.2012.08.003. Epub 2012 Sep 6.
7
Separating Probability and Reversal Learning in a Novel Probabilistic Reversal Learning Task for Mice.在一种用于小鼠的新型概率反转学习任务中分离概率学习和反转学习
Front Behav Neurosci. 2020 Jan 9;13:270. doi: 10.3389/fnbeh.2019.00270. eCollection 2019.
8
Heterogeneity of strategy use in the Iowa gambling task: a comparison of win-stay/lose-shift and reinforcement learning models.策略使用的异质性在爱荷华赌博任务中:赢留输变和强化学习模型的比较。
Psychon Bull Rev. 2013 Apr;20(2):364-71. doi: 10.3758/s13423-012-0324-9.
9
Learning reward frequency over reward probability: A tale of two learning rules.学习奖励频率优于奖励概率:两种学习规则的故事。
Cognition. 2019 Dec;193:104042. doi: 10.1016/j.cognition.2019.104042. Epub 2019 Aug 17.
10
Mice learn to avoid regret.老鼠学会了避免后悔。
PLoS Biol. 2018 Jun 21;16(6):e2005853. doi: 10.1371/journal.pbio.2005853. eCollection 2018 Jun.

引用本文的文献

1
Non-motor symptoms associated with progressive loss of dopaminergic neurons in a mouse model of Parkinson's disease.帕金森病小鼠模型中与多巴胺能神经元渐进性丧失相关的非运动症状。
Front Neurosci. 2024 Apr 30;18:1375265. doi: 10.3389/fnins.2024.1375265. eCollection 2024.

本文引用的文献

1
Ten simple rules for the computational modeling of behavioral data.计算行为数据建模的 10 个简单规则。
Elife. 2019 Nov 26;8:e49547. doi: 10.7554/eLife.49547.
2
The Fat Mass and Obesity-Associated Protein (FTO) Regulates Locomotor Responses to Novelty via D2R Medium Spiny Neurons.脂肪量和肥胖相关蛋白(FTO)通过 D2R 中脑皮层神经元调节对新奇事物的运动反应。
Cell Rep. 2019 Jun 11;27(11):3182-3198.e9. doi: 10.1016/j.celrep.2019.05.037.
3
Selective Effects of the Loss of NMDA or mGluR5 Receptors in the Reward System on Adaptive Decision-Making.
选择性敲除 NMDA 或 mGluR5 受体对奖赏系统适应性决策的影响。
eNeuro. 2018 Oct 5;5(4). doi: 10.1523/ENEURO.0331-18.2018. eCollection 2018 Jul-Aug.
4
An effect of serotonergic stimulation on learning rates for rewards apparent after long intertrial intervals.长的间隔测验后,血清素刺激对奖赏学习率的影响明显。
Nat Commun. 2018 Jun 26;9(1):2477. doi: 10.1038/s41467-018-04840-2.
5
A molecular mechanism for choosing alcohol over an alternative reward.选择酒精而非替代奖励的分子机制。
Science. 2018 Jun 22;360(6395):1321-1326. doi: 10.1126/science.aao1157.
6
Choice for Drug or Natural Reward Engages Largely Overlapping Neuronal Ensembles in the Infralimbic Prefrontal Cortex.药物或自然奖赏的选择主要涉及到边缘前扣带回皮层中的重叠神经元集合。
J Neurosci. 2018 Apr 4;38(14):3507-3519. doi: 10.1523/JNEUROSCI.0026-18.2018. Epub 2018 Feb 26.
7
Trying to make sense of rodents' drug choice behavior.试图理解啮齿动物的药物选择行为。
Prog Neuropsychopharmacol Biol Psychiatry. 2018 Dec 20;87(Pt A):3-10. doi: 10.1016/j.pnpbp.2017.09.027. Epub 2017 Sep 28.
8
Activity patterns of serotonin neurons underlying cognitive flexibility.认知灵活性背后血清素神经元的活动模式。
Elife. 2017 Mar 21;6:e20552. doi: 10.7554/eLife.20552.
9
Working Memory Load Strengthens Reward Prediction Errors.工作记忆负荷增强奖励预测误差。
J Neurosci. 2017 Apr 19;37(16):4332-4342. doi: 10.1523/JNEUROSCI.2700-16.2017. Epub 2017 Mar 20.
10
The neural basis of reversal learning: An updated perspective.反转学习的神经基础:一个更新的视角。
Neuroscience. 2017 Mar 14;345:12-26. doi: 10.1016/j.neuroscience.2016.03.021. Epub 2016 Mar 12.