Suppr超能文献

社会强化学习如何导致亚稳态极化和选民模型。

How social reinforcement learning can lead to metastable polarisation and the voter model.

作者信息

Meylahn Benedikt V, Meylahn Janusz M

机构信息

Korteweg-de Vries Institue for Mathematics, University of Amsterdam, Amsterdam, The Netherlands.

Department of Applied Mathematics, University of Twente, Enschede, The Netherlands.

出版信息

PLoS One. 2024 Dec 17;19(12):e0313951. doi: 10.1371/journal.pone.0313951. eCollection 2024.

Abstract

Previous explanations for the persistence of polarization of opinions have typically included modelling assumptions that predispose the possibility of polarization (i.e., assumptions allowing a pair of agents to drift apart in their opinion such as repulsive interactions or bounded confidence). An exception is a recent simulation study showing that polarization is persistent when agents form their opinions using social reinforcement learning. Our goal is to highlight the usefulness of reinforcement learning in the context of modeling opinion dynamics, but that caution is required when selecting the tools used to study such a model. We show that the polarization observed in the model of the simulation study cannot persist indefinitely, and exhibits consensus asymptotically with probability one. By constructing a link between the reinforcement learning model and the voter model, we argue that the observed polarization is metastable. Finally, we show that a slight modification in the learning process of the agents changes the model from being non-ergodic to being ergodic. Our results show that reinforcement learning may be a powerful method for modelling polarization in opinion dynamics, but that the tools (objects to study such as the stationary distribution, or time to absorption for example) appropriate for analysing such models crucially depend on their properties (such as ergodicity, or transience). These properties are determined by the details of the learning process and may be difficult to identify based solely on simulations.

摘要

以往对观点两极分化持续存在的解释通常包括一些建模假设,这些假设预先设定了两极分化的可能性(即允许一对主体在观点上产生分歧的假设,如排斥性相互作用或有限信心)。一个例外是最近的一项模拟研究表明,当主体使用社会强化学习形成观点时,两极分化会持续存在。我们的目标是强调强化学习在观点动态建模背景下的有用性,但在选择用于研究此类模型的工具时需要谨慎。我们表明,在模拟研究模型中观察到的两极分化不会无限期持续,并且以概率1渐近地表现出一致性。通过在强化学习模型和选民模型之间建立联系,我们认为观察到的两极分化是亚稳态的。最后,我们表明主体学习过程中的轻微修改会使模型从非遍历性变为遍历性。我们的结果表明,强化学习可能是一种用于在观点动态中对两极分化进行建模的强大方法,但适用于分析此类模型的工具(例如用于研究平稳分布或吸收时间等对象)关键取决于它们的属性(如遍历性或暂态性)。这些属性由学习过程的细节决定,仅基于模拟可能难以识别。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fc35/11651571/347e261c5994/pone.0313951.g001.jpg

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验