Department of Computer Science, School of Mathematics, Statistics, and Computer Science, University of Tehran, Tehran, Iran; CERCO UMR 5549, CNRS -Université Toulouse 3, France.
Department of Computer Science, School of Mathematics, Statistics, and Computer Science, University of Tehran, Tehran, Iran.
Neural Netw. 2018 Mar;99:56-67. doi: 10.1016/j.neunet.2017.12.005. Epub 2017 Dec 23.
Previous studies have shown that spike-timing-dependent plasticity (STDP) can be used in spiking neural networks (SNN) to extract visual features of low or intermediate complexity in an unsupervised manner. These studies, however, used relatively shallow architectures, and only one layer was trainable. Another line of research has demonstrated - using rate-based neural networks trained with back-propagation - that having many layers increases the recognition robustness, an approach known as deep learning. We thus designed a deep SNN, comprising several convolutional (trainable with STDP) and pooling layers. We used a temporal coding scheme where the most strongly activated neurons fire first, and less activated neurons fire later or not at all. The network was exposed to natural images. Thanks to STDP, neurons progressively learned features corresponding to prototypical patterns that were both salient and frequent. Only a few tens of examples per category were required and no label was needed. After learning, the complexity of the extracted features increased along the hierarchy, from edge detectors in the first layer to object prototypes in the last layer. Coding was very sparse, with only a few thousands spikes per image, and in some cases the object category could be reasonably well inferred from the activity of a single higher-order neuron. More generally, the activity of a few hundreds of such neurons contained robust category information, as demonstrated using a classifier on Caltech 101, ETH-80, and MNIST databases. We also demonstrate the superiority of STDP over other unsupervised techniques such as random crops (HMAX) or auto-encoders. Taken together, our results suggest that the combination of STDP with latency coding may be a key to understanding the way that the primate visual system learns, its remarkable processing speed and its low energy consumption. These mechanisms are also interesting for artificial vision systems, particularly for hardware solutions.
先前的研究表明,尖峰时间依赖可塑性(STDP)可用于尖峰神经网络(SNN)中,以非监督方式提取低复杂度或中等复杂度的视觉特征。然而,这些研究使用的是相对较浅的架构,且只有一层是可训练的。另一研究方向使用基于速率的神经网络和反向传播训练,表明具有多层可增加识别稳健性,这种方法被称为深度学习。因此,我们设计了一个深度 SNN,包含几个卷积(可通过 STDP 训练)和池化层。我们使用了一种时间编码方案,其中最强激活的神经元首先发射,而较弱激活的神经元稍后发射或根本不发射。该网络暴露于自然图像中。由于 STDP,神经元逐渐学习到对应于原型模式的特征,这些特征既显著又频繁。每个类别只需几十个示例,且无需标签。学习后,提取特征的复杂度沿着层次结构增加,从第一层的边缘检测器到最后一层的对象原型。编码非常稀疏,每张图像只有几千个尖峰,在某些情况下,仅从单个高阶神经元的活动就可以合理地推断出对象类别。更一般地,几百个这样的神经元的活动包含稳健的类别信息,这一点通过在 Caltech 101、ETH-80 和 MNIST 数据库上使用分类器得到证明。我们还证明了 STDP 优于其他无监督技术,例如随机裁剪(HMAX)或自动编码器。总的来说,我们的结果表明,STDP 与延迟编码的结合可能是理解灵长类视觉系统学习方式的关键,其显著的处理速度和低能耗。这些机制对于人工视觉系统也很有趣,特别是对于硬件解决方案。