Department of Mathematical Informatics, The University of Tokyo, 7-3-1 Hongo, Bunkyo, Tokyo 113-8656, Japan.
J Theor Biol. 2011 Jun 7;278(1):55-62. doi: 10.1016/j.jtbi.2011.03.005. Epub 2011 Mar 29.
Humans and other animals can adapt their social behavior in response to environmental cues including the feedback obtained through experience. Nevertheless, the effects of the experience-based learning of players in evolution and maintenance of cooperation in social dilemma games remain relatively unclear. Some previous literature showed that mutual cooperation of learning players is difficult or requires a sophisticated learning model. In the context of the iterated Prisoner's dilemma, we numerically examine the performance of a reinforcement learning model. Our model modifies those of Karandikar et al. (1998), Posch et al. (1999), and Macy and Flache (2002) in which players satisfy if the obtained payoff is larger than a dynamic threshold. We show that players obeying the modified learning mutually cooperate with high probability if the dynamics of threshold is not too fast and the association between the reinforcement signal and the action in the next round is sufficiently strong. The learning players also perform efficiently against the reactive strategy. In evolutionary dynamics, they can invade a population of players adopting simpler but competitive strategies. Our version of the reinforcement learning model does not complicate the previous model and is sufficiently simple yet flexible. It may serve to explore the relationships between learning and evolution in social dilemma situations.
人类和其他动物可以根据环境线索(包括通过经验获得的反馈)来调整其社交行为。然而,玩家基于经验的学习在进化中的作用以及在社会困境游戏中合作的维持仍然相对不清楚。一些先前的文献表明,学习型玩家的相互合作是困难的,或者需要复杂的学习模型。在重复囚徒困境的背景下,我们通过数值方法研究了强化学习模型的性能。我们的模型对 Karandikar 等人(1998 年)、Posch 等人(1999 年)和 Macy 和 Flache(2002 年)的模型进行了修改,在这些模型中,如果获得的回报大于动态阈值,玩家就会得到满足。我们表明,如果阈值的动态变化不是太快,并且下一轮强化信号与动作之间的关联足够强,那么遵循修改后的学习规则的玩家很可能会高度合作。学习型玩家也可以有效地对抗反应性策略。在进化动力学中,他们可以入侵采用简单但具有竞争力策略的玩家群体。我们的强化学习模型版本不会使之前的模型复杂化,它足够简单但灵活。它可以用于探索学习和社会困境情况下的进化之间的关系。