School of Data and Computer Science, Sun Yat-Sen University, China; Department of Computer Science, Technical University of Munich, Germany.
Department of Computer Science, Technical University of Munich, Germany.
Neural Netw. 2020 Jan;121:21-36. doi: 10.1016/j.neunet.2019.05.019. Epub 2019 Jul 9.
Building spiking neural networks (SNNs) based on biological synaptic plasticities holds a promising potential for accomplishing fast and energy-efficient computing, which is beneficial to mobile robotic applications. However, the implementations of SNNs in robotic fields are limited due to the lack of practical training methods. In this paper, we therefore introduce both indirect and direct end-to-end training methods of SNNs for a lane-keeping vehicle. First, we adopt a policy learned using the Deep Q-Learning (DQN) algorithm and then subsequently transfer it to an SNN using supervised learning. Second, we adopt the reward-modulated spike-timing-dependent plasticity (R-STDP) for training SNNs directly, since it combines the advantages of both reinforcement learning and the well-known spike-timing-dependent plasticity (STDP). We examine the proposed approaches in three scenarios in which a robot is controlled to keep within lane markings by using an event-based neuromorphic vision sensor. We further demonstrate the advantages of the R-STDP approach in terms of the lateral localization accuracy and training time steps by comparing them with other three algorithms presented in this paper.
基于生物突触可塑性构建尖峰神经网络(SNN)在实现快速和节能计算方面具有很大的潜力,这有利于移动机器人应用。然而,由于缺乏实用的训练方法,SNN 在机器人领域的应用受到限制。因此,在本文中,我们为车道保持车辆介绍了 SNN 的间接和直接端到端训练方法。首先,我们采用深度 Q 学习(DQN)算法学习的策略,然后使用监督学习将其转换为 SNN。其次,我们采用奖励调制尖峰时间依赖可塑性(R-STDP)直接训练 SNN,因为它结合了强化学习和著名的尖峰时间依赖可塑性(STDP)的优点。我们在三个场景中检验了所提出的方法,其中机器人使用基于事件的神经形态视觉传感器控制以保持在车道标记内。我们通过与本文中提出的其他三种算法进行比较,进一步展示了 R-STDP 方法在横向定位精度和训练时间步长方面的优势。