Gao Haoran, He Junxian, Wang Haibing, Wang Tengxiao, Zhong Zhengqing, Yu Jianyi, Wang Ying, Tian Min, Shi Cong
The School of Microelectronics and Communication Engineering, Chongqing University, Chongqing, China.
State Key Laboratory of Computer Architecture, Institute of Computing Technology, Chinese Academy of Sciences, Beijing, China.
Front Neurosci. 2023 Mar 8;17:1141701. doi: 10.3389/fnins.2023.1141701. eCollection 2023.
Spiking neural networks (SNNs) have attracted intensive attention due to the efficient event-driven computing paradigm. Among SNN training methods, the ANN-to-SNN conversion is usually regarded to achieve state-of-the-art recognition accuracies. However, many existing ANN-to-SNN techniques impose lengthy post-conversion steps like threshold balancing and weight renormalization, to compensate for the inherent behavioral discrepancy between artificial and spiking neurons. In addition, they require a long temporal window to encode and process as many spikes as possible to better approximate the real-valued ANN neurons, leading to a high inference latency. To overcome these challenges, we propose a calcium-gated bipolar leaky integrate and fire (Ca-LIF) spiking neuron model to better approximate the functions of the ReLU neurons widely adopted in ANNs. We also propose a quantization-aware training (QAT)-based framework leveraging an off-the-shelf QAT toolkit for easy ANN-to-SNN conversion, which directly exports the learned ANN weights to SNNs requiring no post-conversion processing. We benchmarked our method on typical deep network structures with varying time-step lengths from 8 to 128. Compared to other research, our converted SNNs reported competitively high-accuracy performance, while enjoying relatively short inference time steps.
脉冲神经网络(SNN)由于其高效的事件驱动计算范式而备受关注。在SNN训练方法中,人工神经网络(ANN)到SNN的转换通常被认为能实现最先进的识别准确率。然而,许多现有的ANN到SNN技术需要冗长的转换后步骤,如阈值平衡和权重归一化,以弥补人工神经元和脉冲神经元之间固有的行为差异。此外,它们需要很长的时间窗口来编码和处理尽可能多的脉冲,以便更好地逼近实值的ANN神经元,从而导致较高的推理延迟。为了克服这些挑战,我们提出了一种钙门控双极泄漏积分发放(Ca-LIF)脉冲神经元模型,以更好地逼近ANN中广泛采用的ReLU神经元的功能。我们还提出了一个基于量化感知训练(QAT)的框架,利用现成的QAT工具包实现简单的ANN到SNN转换,该框架直接将学习到的ANN权重导出到SNN,无需进行转换后处理。我们在时间步长从8到128变化的典型深度网络结构上对我们的方法进行了基准测试。与其他研究相比,我们转换后的SNN在具有较高准确率的同时,推理时间步长相对较短。