IEEE Trans Neural Netw Learn Syst. 2019 Jul;30(7):2173-2187. doi: 10.1109/TNNLS.2018.2878002. Epub 2018 Nov 20.
Memristive crossbars have become a popular means for realizing unsupervised and supervised learning techniques. In previous neuromorphic architectures with leaky integrate-and-fire neurons, the crossbar itself has been separated from the neuron capacitors to preserve mathematical rigor. In this paper, we sought to design a simplified sparse coding circuit without this restriction, resulting in a fast circuit that approximated a sparse coding operation at a minimal loss in accuracy. We showed that connecting the neurons directly to the crossbar resulted in a more energy-efficient sparse coding architecture and alleviated the need to prenormalize receptive fields. This paper provides derivations for the design of such a network, named the simple spiking locally competitive algorithm, as well as CMOS designs and results on the CIFAR and MNIST data sets. Compared to a nonspiking, nonapproximate model which scored 33% on CIFAR-10 with a single-layer classifier, this hardware scored 32% accuracy. When used with a state-of-the-art deep learning classifier, the nonspiking model achieved 82% and our simplified, spiking model achieved 80% while compressing the input data by 92%. Compared to a previously proposed spiking model, our proposed hardware consumed 99% less energy to do the same work at 21 × the throughput. Accuracy held out with online learning to a write variance of 3%, suitable for the often reported 4-bit resolution required for neuromorphic algorithms, with offline learning to a write variance of 27%, and with read variance to 40%. The proposed architecture's excellent accuracy, throughput, and significantly lower energy usage demonstrate the utility of our innovations.
忆阻器交叉点已成为实现无监督和监督学习技术的一种流行手段。在以前具有漏电流积分和放电神经元的神经形态架构中,交叉点本身已经与神经元电容器分离,以保持数学严谨性。在本文中,我们试图设计一种没有这种限制的简化稀疏编码电路,从而得到一个快速电路,在最小精度损失的情况下近似稀疏编码操作。我们表明,将神经元直接连接到交叉点会导致更节能的稀疏编码架构,并减轻了对预归一化感受野的需求。本文提供了这种名为简单脉冲局部竞争算法的网络设计推导,以及 CMOS 设计和 CIFAR 和 MNIST 数据集上的结果。与使用单层分类器在 CIFAR-10 上得分为 33%的非脉冲、非近似模型相比,该硬件的准确率为 32%。当与最先进的深度学习分类器一起使用时,非脉冲模型的准确率为 82%,而我们简化的脉冲模型的准确率为 80%,同时将输入数据压缩了 92%。与之前提出的脉冲模型相比,我们提出的硬件在 21 倍的吞吐量下,以 99%的能耗完成相同的工作。在线学习的准确率保持在 3%的写入方差,适合神经形态算法通常需要的 4 位分辨率,离线学习的写入方差为 27%,读取方差为 40%。所提出的架构具有出色的准确性、吞吐量和显著降低的能耗,证明了我们创新的实用性。