Suppr超能文献

作为强化物“持续时间”函数的选择:从概率学习到并发强化

Choice as a function of reinforcer "hold": from probability learning to concurrent reinforcement.

作者信息

Jensen Greg, Neuringer Allen

机构信息

Psychology Department, Reed College, Portland, OR 97202, USA.

出版信息

J Exp Psychol Anim Behav Process. 2008 Oct;34(4):437-60. doi: 10.1037/0097-7403.34.4.437.

Abstract

Two procedures commonly used to study choice are concurrent reinforcement and probability learning. Under concurrent-reinforcement procedures, once a reinforcer is scheduled, it remains available indefinitely until collected. Therefore reinforcement becomes increasingly likely with passage of time or responses on other operanda. Under probability learning, reinforcer probabilities are constant and independent of passage of time or responses. Therefore a particular reinforcer is gained or not, on the basis of a single response, and potential reinforcers are not retained, as when betting at a roulette wheel. In the "real" world, continued availability of reinforcers often lies between these two extremes, with potential reinforcers being lost owing to competition, maturation, decay, and random scatter. The authors parametrically manipulated the likelihood of continued reinforcer availability, defined as hold, and examined the effects on pigeons' choices. Choices varied as power functions of obtained reinforcers under all values of hold. Stochastic models provided generally good descriptions of choice emissions with deviations from stochasticity systematically related to hold. Thus, a single set of principles accounted for choices across hold values that represent a wide range of real-world conditions.

摘要

两种常用于研究选择的程序是并发强化和概率学习。在并发强化程序中,一旦安排了强化物,它会一直可用,直到被获取。因此,随着时间的推移或在其他操作上的反应,强化变得越来越有可能。在概率学习中,强化物的概率是恒定的,与时间的推移或反应无关。因此,基于单次反应,特定的强化物要么获得,要么未获得,并且潜在的强化物不会像在轮盘赌下注时那样被保留。在“现实”世界中,强化物的持续可用性通常介于这两个极端之间,由于竞争、成熟、衰退和随机分散,潜在的强化物会丢失。作者对持续强化物可用性的可能性(定义为持有)进行了参数化操作,并研究了其对鸽子选择的影响。在所有持有值下,选择随着获得的强化物的幂函数而变化。随机模型通常能很好地描述选择发出情况,与随机性的偏差与持有系统相关。因此,一组单一的原则解释了跨越代表广泛现实世界条件的持有值的选择。

相似文献

1
Choice as a function of reinforcer "hold": from probability learning to concurrent reinforcement.
J Exp Psychol Anim Behav Process. 2008 Oct;34(4):437-60. doi: 10.1037/0097-7403.34.4.437.
2
Examining the discriminative and strengthening effects of reinforcers in concurrent schedules.
J Exp Anal Behav. 2011 Sep;96(2):227-41. doi: 10.1901/jeab.2011.96-227.
3
Local preference in concurrent schedules: the effects of reinforcer sequences.
J Exp Anal Behav. 2005 Jul;84(1):37-64. doi: 10.1901/jeab.2005.114-04.
4
Barycentric extension of generalized matching.
J Exp Anal Behav. 2009 Sep;92(2):139-59. doi: 10.1901/jeab.2009.92-139.
5
Being there on time: Reinforcer effects on timing and locating.
J Exp Anal Behav. 2020 Mar;113(2):340-362. doi: 10.1002/jeab.581. Epub 2020 Jan 28.
6
Effects of time between trials on rats' and pigeons' choices with probabilistic delayed reinforcers.
J Exp Anal Behav. 2011 Jan;95(1):41-56. doi: 10.1901/jeab.2011.95-41.
7
Signaled reinforcement: Effects of signal reliability on choice between signaled and unsignaled alternatives.
Behav Processes. 2020 May;174:104088. doi: 10.1016/j.beproc.2020.104088. Epub 2020 Feb 21.
8
Species differences between rats and pigeons in choices with probabilistic and delayed reinforcers.
Behav Processes. 2007 Jun;75(2):220-4. doi: 10.1016/j.beproc.2007.02.004. Epub 2007 Feb 8.
9
Choosing among multiple alternatives: Relative and overall reinforcer rates.
J Exp Anal Behav. 2017 Sep;108(2):204-222. doi: 10.1002/jeab.269. Epub 2017 Jul 31.
10
Pigeons' discounting of probabilistic and delayed reinforcers.
J Exp Anal Behav. 2010 Sep;94(2):113-23. doi: 10.1901/jeab.2010.94-113.

引用本文的文献

1
The active time model of concurrent choice.
PLoS One. 2024 May 21;19(5):e0301173. doi: 10.1371/journal.pone.0301173. eCollection 2024.
2
Dynamics of pre- and post-choice behaviour: rats approximate optimal strategy in a discrete-trial decision task.
Proc Biol Sci. 2015 Mar 22;282(1803):20142963. doi: 10.1098/rspb.2014.2963.
3
Information: theory, brain, and behavior.
J Exp Anal Behav. 2013 Nov;100(3):408-31. doi: 10.1002/jeab.49. Epub 2013 Oct 4.
4
Reinforcement and induction of operant variability.
Behav Anal. 2012 Fall;35(2):229-35. doi: 10.1007/BF03392281.
5
Barycentric extension of generalized matching.
J Exp Anal Behav. 2009 Sep;92(2):139-59. doi: 10.1901/jeab.2009.92-139.

本文引用的文献

1
The role of shifting in choice behavior of pigeons on a two-armed bandit.
Behav Processes. 1990 Jun;21(2-3):157-78. doi: 10.1016/0376-6357(90)90022-8.
2
Is matching innate?
J Exp Anal Behav. 2007 Mar;87(2):161-99. doi: 10.1901/jeab.2007.92-05.
3
Maternal nutrition and four-alternative choice.
J Exp Anal Behav. 2007 Jan;87(1):51-62. doi: 10.1901/jeab.2007.12-06.
4
A local model of concurrent performance.
J Exp Anal Behav. 1999 Jan;71(1):57-74. doi: 10.1901/jeab.1999.71-57.
5
Behavior analysis and decision making.
J Exp Anal Behav. 1998 May;69(3):355-64. doi: 10.1901/jeab.1998.69-355.
6
Concurrent variable-ratio schedules: Implications for the generalized matching law.
J Exp Anal Behav. 1988 Jul;50(1):55-64. doi: 10.1901/jeab.1988.50-55.
7
Competitive fixed-interval performance in humans.
J Exp Anal Behav. 1987 Mar;47(2):145-58. doi: 10.1901/jeab.1987.47-145.
8
Melloration and maximization of reinforcement minus costs of behavior.
J Exp Anal Behav. 1984 Jul;42(1):113-26. doi: 10.1901/jeab.1984.42-113.
9
Optimal choice.
J Exp Anal Behav. 1981 May;35(3):397-412. doi: 10.1901/jeab.1981.35-397.
10
A Markov model description of changeover probabilities on concurrent variable-interval schedules.
J Exp Anal Behav. 1979 Jan;31(1):41-51. doi: 10.1901/jeab.1979.31-41.

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验