• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

进一步研究动物如何随着时间推移权衡来自奖励源的相互冲突的信息。

Further examining how animals weigh conflicting information about reward sources over time.

作者信息

Van Allsburg Jack, Shahan Timothy A

机构信息

Department of Psychology, Utah State University, Logan, UT, USA.

出版信息

Anim Cogn. 2025 Jul 30;28(1):74. doi: 10.1007/s10071-025-01982-x.

DOI:10.1007/s10071-025-01982-x
PMID:40736588
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12310909/
Abstract

Spontaneous recovery of choice is a behavioral phenomenon where a delay period (without new experience) elicits the recovery of a preference consistent with a previous distribution of rewards, rather than the most recently experienced distribution of rewards. On short timescales (< 48 h), the occurrence of spontaneous recovery of choice has been effectively predicted by the Temporal Weighting Rule. However, previous study of this phenomenon over longer timescales (> 48 h) has found results inconsistent with model predictions. The present experiments investigated three potential explanations for these results: (1.) whether time's passage alone causes animals to revert to random exploratory behavior; (2.) whether time's effect on behavior is moderated by experience of volatility in rewards during training; and (3.) whether a drift toward random exploratory behavior produced by time's passage can be distinguished from the effect of spontaneous recovery of choice. Subjects experienced varied reward conditions in a concurrent choice procedure before preference between options was evaluated at various test delays. Obtained results ruled out these first two explanations, but were inconclusive in distinguishing the effects of a drift toward random exploratory behavior from the effect of spontaneous recovery of choice. Limitations and directions for further investigation are discussed.

摘要

选择的自发恢复是一种行为现象,即在一段延迟期(无新经验)后,会引发与先前奖励分布一致的偏好恢复,而非最近经历的奖励分布。在短时间尺度(<48小时)上,选择的自发恢复的发生已通过时间加权规则得到有效预测。然而,此前在更长时间尺度(>48小时)上对该现象的研究发现结果与模型预测不一致。本实验探究了对这些结果的三种潜在解释:(1)时间的流逝本身是否会导致动物恢复到随机探索行为;(2)时间对行为的影响是否会受到训练期间奖励波动性经验的调节;(3)时间流逝产生的向随机探索行为的漂移是否能与选择的自发恢复的影响区分开来。在并行选择程序中,受试者经历了不同的奖励条件,然后在不同的测试延迟下评估选项之间的偏好。所得结果排除了前两种解释,但在区分向随机探索行为的漂移的影响与选择的自发恢复的影响方面尚无定论。讨论了局限性和进一步研究的方向。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dfeb/12310909/f8d70aa88b89/10071_2025_1982_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dfeb/12310909/4685cfc68a09/10071_2025_1982_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dfeb/12310909/8f63f0fe2c5d/10071_2025_1982_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dfeb/12310909/f8d70aa88b89/10071_2025_1982_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dfeb/12310909/4685cfc68a09/10071_2025_1982_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dfeb/12310909/8f63f0fe2c5d/10071_2025_1982_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dfeb/12310909/f8d70aa88b89/10071_2025_1982_Fig3_HTML.jpg

相似文献

1
Further examining how animals weigh conflicting information about reward sources over time.进一步研究动物如何随着时间推移权衡来自奖励源的相互冲突的信息。
Anim Cogn. 2025 Jul 30;28(1):74. doi: 10.1007/s10071-025-01982-x.
2
The Black Book of Psychotropic Dosing and Monitoring.《精神药物剂量与监测黑皮书》
Psychopharmacol Bull. 2024 Jul 8;54(3):8-59.
3
Sexual Harassment and Prevention Training性骚扰与预防培训
4
Short-Term Memory Impairment短期记忆障碍
5
A rapid and systematic review of the clinical effectiveness and cost-effectiveness of paclitaxel, docetaxel, gemcitabine and vinorelbine in non-small-cell lung cancer.对紫杉醇、多西他赛、吉西他滨和长春瑞滨在非小细胞肺癌中的临床疗效和成本效益进行的快速系统评价。
Health Technol Assess. 2001;5(32):1-195. doi: 10.3310/hta5320.
6
Comparison of Two Modern Survival Prediction Tools, SORG-MLA and METSSS, in Patients With Symptomatic Long-bone Metastases Who Underwent Local Treatment With Surgery Followed by Radiotherapy and With Radiotherapy Alone.两种现代生存预测工具 SORG-MLA 和 METSSS 在接受手术联合放疗和单纯放疗治疗有症状长骨转移患者中的比较。
Clin Orthop Relat Res. 2024 Dec 1;482(12):2193-2208. doi: 10.1097/CORR.0000000000003185. Epub 2024 Jul 23.
7
Maternal and neonatal outcomes of elective induction of labor.择期引产的母婴结局
Evid Rep Technol Assess (Full Rep). 2009 Mar(176):1-257.
8
The quantity, quality and findings of network meta-analyses evaluating the effectiveness of GLP-1 RAs for weight loss: a scoping review.评估胰高血糖素样肽-1受体激动剂(GLP-1 RAs)减肥效果的网状Meta分析的数量、质量及结果:一项范围综述
Health Technol Assess. 2025 Jun 25:1-73. doi: 10.3310/SKHT8119.
9
Intravenous magnesium sulphate and sotalol for prevention of atrial fibrillation after coronary artery bypass surgery: a systematic review and economic evaluation.静脉注射硫酸镁和索他洛尔预防冠状动脉搭桥术后房颤:系统评价与经济学评估
Health Technol Assess. 2008 Jun;12(28):iii-iv, ix-95. doi: 10.3310/hta12280.
10
Melatonin versus midazolam in the premedication of anxious children attending for elective surgery under general anaesthesia: the MAGIC non-inferiority RCT.褪黑素与咪达唑仑用于择期全身麻醉手术患儿术前用药的比较:MAGIC非劣效性随机对照试验
Health Technol Assess. 2025 Jul;29(29):1-25. doi: 10.3310/CWKF1987.

本文引用的文献

1
How do animals weigh conflicting information about reward sources over time? Comparing dynamic averaging models.动物如何随着时间的推移权衡关于奖励来源的冲突信息?比较动态平均模型。
Anim Cogn. 2024 Mar 2;27(1):11. doi: 10.1007/s10071-024-01840-2.
2
Value-free reinforcement learning: policy optimization as a minimal model of operant behavior.无价值强化学习:作为操作性行为最小模型的策略优化
Curr Opin Behav Sci. 2021 Oct;41:114-121. doi: 10.1016/j.cobeha.2021.04.020. Epub 2021 May 28.
3
A model for learning based on the joint estimation of stochasticity and volatility.
基于随机波动联合估计的学习模型。
Nat Commun. 2021 Nov 15;12(1):6587. doi: 10.1038/s41467-021-26731-9.
4
Uncertainty and Exploration.不确定性与探索。
Decision (Wash D C ). 2019 Jul;6(3):277-286. doi: 10.1037/dec0000101. Epub 2018 Oct 1.
5
Matching Behaviours and Rewards.匹配行为和奖励。
Trends Cogn Sci. 2021 May;25(5):403-415. doi: 10.1016/j.tics.2021.01.011. Epub 2021 Feb 18.
6
Relapse: An introduction.复发:引言
J Exp Anal Behav. 2020 Jan;113(1):8-14. doi: 10.1002/jeab.578. Epub 2020 Jan 3.
7
Deviation from the matching law reflects an optimal strategy involving learning over multiple timescales.偏离匹配律反映了一种涉及多个时间尺度的学习的最优策略。
Nat Commun. 2019 Apr 1;10(1):1466. doi: 10.1038/s41467-019-09388-3.
8
Joint modeling of reaction times and choice improves parameter identifiability in reinforcement learning models.联合建模反应时间和选择可提高强化学习模型的参数可识别性。
J Neurosci Methods. 2019 Apr 1;317:37-44. doi: 10.1016/j.jneumeth.2019.01.006. Epub 2019 Jan 18.
9
Toward a contemporary quantitative model of punishment.迈向当代惩罚定量模型。
J Exp Anal Behav. 2018 Mar;109(2):336-348. doi: 10.1002/jeab.317. Epub 2018 Mar 6.
10
A Primer on Foraging and the Explore/Exploit Trade-Off for Psychiatry Research.精神病学研究中的觅食及探索/利用权衡入门
Neuropsychopharmacology. 2017 Sep;42(10):1931-1939. doi: 10.1038/npp.2017.108. Epub 2017 May 29.