Liu Liu, Xu Zhifei
College of Business Administration, Capital University of Economics and Business, Beijing, China.
School of Science and Engineering, Chinese University of Hong Kong - Shenzhen, Shenzhen, Guangdong, China.
PLoS One. 2025 May 15;20(5):e0320777. doi: 10.1371/journal.pone.0320777. eCollection 2025.
This research explores the potential of combining Meta Reinforcement Learning (MRL) with Spike-Timing-Dependent Plasticity (STDP) to enhance the performance and adaptability of AI agents in Atari game settings. Our methodology leverages MRL to swiftly adjust agent strategies across a range of games, while STDP fine-tunes synaptic weights based on neuronal spike timings, which in turn improves learning efficiency and decision-making under changing conditions. A series of experiments were conducted with standard Atari games to compare the hybrid MRL-STDP model against baseline models using traditional reinforcement learning techniques like Q-learning and Deep Q-Networks. Various performance metrics, including learning speed, adaptability, and cross-game generalization, were evaluated. The results show that the MRL-STDP approach significantly accelerates the agent's ability to reach competitive performance levels, with a 40% boost in learning efficiency and a 35% increase in adaptability over conventional models.
本研究探讨了将元强化学习(MRL)与尖峰时间依赖可塑性(STDP)相结合的潜力,以提高人工智能智能体在雅达利游戏环境中的性能和适应性。我们的方法利用MRL在一系列游戏中迅速调整智能体策略,而STDP则根据神经元尖峰时间微调突触权重,进而提高学习效率并在变化的条件下改善决策。我们使用标准雅达利游戏进行了一系列实验,以将混合MRL-STDP模型与使用传统强化学习技术(如Q学习和深度Q网络)的基线模型进行比较。评估了各种性能指标,包括学习速度、适应性和跨游戏泛化能力。结果表明,MRL-STDP方法显著加快了智能体达到竞争性能水平的能力,与传统模型相比,学习效率提高了40%,适应性提高了35%。