Xiao Mingqing, Meng Qingyan, Zhang Zongpeng, Wang Yisen, Lin Zhouchen
National Key Laboratory of General Artificial Intelligence, School of Intelligence Science and Technology, Peking University, China.
The Chinese University of Hong Kong, Shenzhen, China; Shenzhen Research Institute of Big Data, Shenzhen 518115, China.
Neural Netw. 2023 Apr;161:9-24. doi: 10.1016/j.neunet.2023.01.026. Epub 2023 Jan 24.
Spiking neural networks (SNNs) with event-based computation are promising brain-inspired models for energy-efficient applications on neuromorphic hardware. However, most supervised SNN training methods, such as conversion from artificial neural networks or direct training with surrogate gradients, require complex computation rather than spike-based operations of spiking neurons during training. In this paper, we study spike-based implicit differentiation on the equilibrium state (SPIDE) that extends the recently proposed training method, implicit differentiation on the equilibrium state (IDE), for supervised learning with purely spike-based computation, which demonstrates the potential for energy-efficient training of SNNs. Specifically, we introduce ternary spiking neuron couples and prove that implicit differentiation can be solved by spikes based on this design, so the whole training procedure, including both forward and backward passes, is made as event-driven spike computation, and weights are updated locally with two-stage average firing rates. Then we propose to modify the reset membrane potential to reduce the approximation error of spikes. With these key components, we can train SNNs with flexible structures in a small number of time steps and with firing sparsity during training, and the theoretical estimation of energy costs demonstrates the potential for high efficiency. Meanwhile, experiments show that even with these constraints, our trained models can still achieve competitive results on MNIST, CIFAR-10, CIFAR-100, and CIFAR10-DVS.
具有基于事件计算的脉冲神经网络(SNN)是有望用于神经形态硬件上节能应用的受大脑启发的模型。然而,大多数监督式SNN训练方法,例如从人工神经网络转换或使用替代梯度直接训练,在训练期间需要复杂的计算而非脉冲神经元基于脉冲的操作。在本文中,我们研究了基于脉冲的平衡状态隐式微分(SPIDE),它扩展了最近提出的用于监督学习的平衡状态隐式微分(IDE)训练方法,采用纯基于脉冲的计算,这展示了SNN节能训练的潜力。具体而言,我们引入三元脉冲神经元对,并证明基于此设计可以通过脉冲解决隐式微分,因此整个训练过程,包括前向和反向传播,都作为事件驱动的脉冲计算,并且权重通过两阶段平均发放率进行局部更新。然后我们提议修改重置膜电位以减少脉冲的近似误差。有了这些关键组件,我们可以在少量时间步长内训练具有灵活结构的SNN,并且在训练期间具有发放稀疏性,并且能量成本的理论估计证明了高效率的潜力。同时,实验表明,即使有这些限制,我们训练的模型在MNIST、CIFAR-10、CIFAR-100和CIFAR10-DVS上仍然可以取得有竞争力的结果。