Suppr超能文献

具有迭代囚徒困境中动态期望水平的强化学习模型的数值分析。

Numerical analysis of a reinforcement learning model with the dynamic aspiration level in the iterated Prisoner's dilemma.

机构信息

Department of Mathematical Informatics, The University of Tokyo, 7-3-1 Hongo, Bunkyo, Tokyo 113-8656, Japan.

出版信息

J Theor Biol. 2011 Jun 7;278(1):55-62. doi: 10.1016/j.jtbi.2011.03.005. Epub 2011 Mar 29.

Abstract

Humans and other animals can adapt their social behavior in response to environmental cues including the feedback obtained through experience. Nevertheless, the effects of the experience-based learning of players in evolution and maintenance of cooperation in social dilemma games remain relatively unclear. Some previous literature showed that mutual cooperation of learning players is difficult or requires a sophisticated learning model. In the context of the iterated Prisoner's dilemma, we numerically examine the performance of a reinforcement learning model. Our model modifies those of Karandikar et al. (1998), Posch et al. (1999), and Macy and Flache (2002) in which players satisfy if the obtained payoff is larger than a dynamic threshold. We show that players obeying the modified learning mutually cooperate with high probability if the dynamics of threshold is not too fast and the association between the reinforcement signal and the action in the next round is sufficiently strong. The learning players also perform efficiently against the reactive strategy. In evolutionary dynamics, they can invade a population of players adopting simpler but competitive strategies. Our version of the reinforcement learning model does not complicate the previous model and is sufficiently simple yet flexible. It may serve to explore the relationships between learning and evolution in social dilemma situations.

摘要

人类和其他动物可以根据环境线索(包括通过经验获得的反馈)来调整其社交行为。然而,玩家基于经验的学习在进化中的作用以及在社会困境游戏中合作的维持仍然相对不清楚。一些先前的文献表明,学习型玩家的相互合作是困难的,或者需要复杂的学习模型。在重复囚徒困境的背景下,我们通过数值方法研究了强化学习模型的性能。我们的模型对 Karandikar 等人(1998 年)、Posch 等人(1999 年)和 Macy 和 Flache(2002 年)的模型进行了修改,在这些模型中,如果获得的回报大于动态阈值,玩家就会得到满足。我们表明,如果阈值的动态变化不是太快,并且下一轮强化信号与动作之间的关联足够强,那么遵循修改后的学习规则的玩家很可能会高度合作。学习型玩家也可以有效地对抗反应性策略。在进化动力学中,他们可以入侵采用简单但具有竞争力策略的玩家群体。我们的强化学习模型版本不会使之前的模型复杂化,它足够简单但灵活。它可以用于探索学习和社会困境情况下的进化之间的关系。

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验