Okinawa Institute of Science and Technology Graduate University, Onna, Okinawa, Japan.
J Neurophysiol. 2021 Jun 1;125(6):2034-2037. doi: 10.1152/jn.00712.2020. Epub 2021 Apr 28.
A common pitfall of current reinforcement learning agents implemented in computational models is in their inadaptability postoptimization. Najarro and Risi [Najarro E, Risi S. . 2020: 20719-20731, 2020] demonstrate how such adaptability may be salvaged in artificial feed-forward networks by optimizing coefficients of classic Hebbian rules to dynamically control the networks' weights instead of optimizing the weights directly. Although such models fail to capture many important neurophysiological details, allying the fields of neuroscience and artificial intelligence in this way bears many fruits for both fields, especially when computational models engage with topics with a rich history in neuroscience such as Hebbian plasticity.
当前在计算模型中实现的强化学习代理的一个常见陷阱是优化后不适应。Najarro 和 Risi [Najarro E,Risi S.. 2020:20719-20731,2020] 演示了如何通过优化经典赫布规则的系数来动态控制网络的权重,而不是直接优化权重,从而在人工前馈网络中挽救这种适应性。尽管这些模型未能捕捉到许多重要的神经生理学细节,但以这种方式将神经科学和人工智能领域结合起来,对这两个领域都有很多好处,特别是当计算模型涉及神经科学中具有丰富历史的主题,如赫布可塑性时。