• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

SpQuant-SNN:具有稀疏激活的超低精度膜电位开启了片上脉冲神经网络应用的潜力。

SpQuant-SNN: ultra-low precision membrane potential with sparse activations unlock the potential of on-device spiking neural networks applications.

作者信息

Hasssan Ahmed, Meng Jian, Anupreetham Anupreetham, Seo Jae-Sun

机构信息

School of Electrical and Computer Engineering, Cornell Tech, New York, NY, United States.

出版信息

Front Neurosci. 2024 Sep 4;18:1440000. doi: 10.3389/fnins.2024.1440000. eCollection 2024.

DOI:10.3389/fnins.2024.1440000
PMID:39296710
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11408473/
Abstract

Spiking neural networks (SNNs) have received increasing attention due to their high biological plausibility and energy efficiency. The binary spike-based information propagation enables efficient sparse computation in event-based and static computer vision applications. However, the weight precision and especially the membrane potential precision remain as high-precision values (e.g., 32 bits) in state-of-the-art SNN algorithms. Each neuron in an SNN stores the membrane potential over time and typically updates its value in every time step. Such frequent read/write operations of high-precision membrane potential incur storage and memory access overhead in SNNs, which undermines the SNNs' compatibility with resource-constrained hardware. To resolve this inefficiency, prior works have explored the time step reduction and low-precision representation of membrane potential at a limited scale and reported significant accuracy drops. Furthermore, while recent advances in on-device AI present pruning and quantization optimization with different architectures and datasets, simultaneous pruning with quantization is highly under-explored in SNNs. In this work, we present , a fully-quantized spiking neural network with , enabling the end-to-end low precision with significantly reduced operations on SNN. First, we propose an integer-only quantization scheme for the membrane potential with a stacked surrogate gradient function, a simple-yet-effective method that enables the smooth learning process of quantized SNN training. Second, we implement spatial-channel pruning with membrane potential prior, toward reducing the layer-wise computational complexity, and floating-point operations (FLOPs) in SNNs. Finally, to further improve the accuracy of low-precision and sparse SNN, we propose a self-adaptive learnable potential threshold for SNN training. Equipped with high biological adaptiveness, minimal computations, and memory utilization, SpQuant-SNN achieves state-of-the-art performance across multiple SNN models for both event-based and static image datasets, including both image classification and object detection tasks. The proposed SpQuant-SNN achieved up to 13× memory reduction and >4.7× FLOPs reduction with < 1.8% accuracy degradation for both classification and object detection tasks, compared to the SOTA baseline.

摘要

脉冲神经网络(SNN)因其高度的生物合理性和能源效率而受到越来越多的关注。基于二进制脉冲的信息传播在基于事件的和静态的计算机视觉应用中实现了高效的稀疏计算。然而,在当前最先进的SNN算法中,权重精度,尤其是膜电位精度仍保持为高精度值(例如32位)。SNN中的每个神经元会随时间存储膜电位,并通常在每个时间步更新其值。这种对高精度膜电位的频繁读/写操作在SNN中会产生存储和内存访问开销,这削弱了SNN与资源受限硬件的兼容性。为了解决这种低效率问题,先前的工作在有限规模上探索了减少时间步长和膜电位的低精度表示,并报告了显著的精度下降。此外,虽然设备端人工智能的最新进展提出了针对不同架构和数据集的剪枝和量化优化,但在SNN中同时进行剪枝和量化的研究还非常少。在这项工作中,我们提出了SpQuant-SNN,一种完全量化的脉冲神经网络,通过显著减少SNN上的操作实现了端到端的低精度。首先,我们为膜电位提出了一种仅整数的量化方案,采用堆叠替代梯度函数,这是一种简单而有效的方法,能够实现量化SNN训练的平滑学习过程。其次,我们利用膜电位先验实现空间通道剪枝,以降低SNN中层级的计算复杂度和浮点运算次数(FLOPs)。最后,为了进一步提高低精度和稀疏SNN的精度,我们为SNN训练提出了一种自适应可学习的电位阈值。SpQuant-SNN具有高度的生物适应性、最少的计算量和内存利用率,在基于事件的和静态图像数据集的多个SNN模型上实现了最先进的性能,包括图像分类和目标检测任务。与最先进的基线相比,所提出的SpQuant-SNN在分类和目标检测任务中实现了高达13倍的内存减少和>4.7倍的FLOPs减少,精度下降<1.8%。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f255/11408473/78dd4f1a4f03/fnins-18-1440000-g0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f255/11408473/86c9beaad177/fnins-18-1440000-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f255/11408473/33691d7832e6/fnins-18-1440000-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f255/11408473/3db95840b94e/fnins-18-1440000-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f255/11408473/6793d0ee7f94/fnins-18-1440000-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f255/11408473/78dd4f1a4f03/fnins-18-1440000-g0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f255/11408473/86c9beaad177/fnins-18-1440000-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f255/11408473/33691d7832e6/fnins-18-1440000-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f255/11408473/3db95840b94e/fnins-18-1440000-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f255/11408473/6793d0ee7f94/fnins-18-1440000-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f255/11408473/78dd4f1a4f03/fnins-18-1440000-g0005.jpg

相似文献

1
SpQuant-SNN: ultra-low precision membrane potential with sparse activations unlock the potential of on-device spiking neural networks applications.SpQuant-SNN:具有稀疏激活的超低精度膜电位开启了片上脉冲神经网络应用的潜力。
Front Neurosci. 2024 Sep 4;18:1440000. doi: 10.3389/fnins.2024.1440000. eCollection 2024.
2
Fast-SNN: Fast Spiking Neural Network by Converting Quantized ANN.快速脉冲神经网络:通过量化人工神经网络转换实现的快速脉冲神经网络
IEEE Trans Pattern Anal Mach Intell. 2023 Dec;45(12):14546-14562. doi: 10.1109/TPAMI.2023.3275769. Epub 2023 Nov 3.
3
Neuron pruning in temporal domain for energy efficient SNN processor design.用于高能效脉冲神经网络(SNN)处理器设计的时域神经元修剪
Front Neurosci. 2023 Nov 30;17:1285914. doi: 10.3389/fnins.2023.1285914. eCollection 2023.
4
Optimizing Deeper Spiking Neural Networks for Dynamic Vision Sensing.深度尖峰神经网络在动态视觉传感中的优化。
Neural Netw. 2021 Dec;144:686-698. doi: 10.1016/j.neunet.2021.09.022. Epub 2021 Oct 5.
5
SSTDP: Supervised Spike Timing Dependent Plasticity for Efficient Spiking Neural Network Training.SSTDP:用于高效脉冲神经网络训练的监督式脉冲时间依赖可塑性
Front Neurosci. 2021 Nov 4;15:756876. doi: 10.3389/fnins.2021.756876. eCollection 2021.
6
ALBSNN: ultra-low latency adaptive local binary spiking neural network with accuracy loss estimator.ALBSNN:具有精度损失估计器的超低延迟自适应局部二值脉冲神经网络
Front Neurosci. 2023 Sep 13;17:1225871. doi: 10.3389/fnins.2023.1225871. eCollection 2023.
7
Trainable quantization for Speedy Spiking Neural Networks.用于快速脉冲神经网络的可训练量化
Front Neurosci. 2023 Mar 3;17:1154241. doi: 10.3389/fnins.2023.1154241. eCollection 2023.
8
SPIDEN: deep Spiking Neural Networks for efficient image denoising.SPIDEN:用于高效图像去噪的深度脉冲神经网络。
Front Neurosci. 2023 Aug 11;17:1224457. doi: 10.3389/fnins.2023.1224457. eCollection 2023.
9
DIET-SNN: A Low-Latency Spiking Neural Network With Direct Input Encoding and Leakage and Threshold Optimization.DIET-SNN:一种具有直接输入编码以及泄漏和阈值优化的低延迟脉冲神经网络。
IEEE Trans Neural Netw Learn Syst. 2023 Jun;34(6):3174-3182. doi: 10.1109/TNNLS.2021.3111897. Epub 2023 Jun 1.
10
ACE-SNN: Algorithm-Hardware Co-design of Energy-Efficient & Low-Latency Deep Spiking Neural Networks for 3D Image Recognition.ACE-SNN:用于3D图像识别的高能效与低延迟深度脉冲神经网络的算法-硬件协同设计
Front Neurosci. 2022 Apr 7;16:815258. doi: 10.3389/fnins.2022.815258. eCollection 2022.

本文引用的文献

1
Trainable quantization for Speedy Spiking Neural Networks.用于快速脉冲神经网络的可训练量化
Front Neurosci. 2023 Mar 3;17:1154241. doi: 10.3389/fnins.2023.1154241. eCollection 2023.
2
Quantization Framework for Fast Spiking Neural Networks.快速脉冲神经网络的量化框架
Front Neurosci. 2022 Jul 19;16:918793. doi: 10.3389/fnins.2022.918793. eCollection 2022.
3
Backpropagation with biologically plausible spatiotemporal adjustment for training deep spiking neural networks.用于训练深度脉冲神经网络的具有生物合理时空调整的反向传播。
Patterns (N Y). 2022 Jun 2;3(6):100522. doi: 10.1016/j.patter.2022.100522. eCollection 2022 Jun 10.
4
Event-Based Vision: A Survey.基于事件的视觉:综述。
IEEE Trans Pattern Anal Mach Intell. 2022 Jan;44(1):154-180. doi: 10.1109/TPAMI.2020.3008413. Epub 2021 Dec 7.
5
Going Deeper in Spiking Neural Networks: VGG and Residual Architectures.深入探索脉冲神经网络:VGG和残差架构。
Front Neurosci. 2019 Mar 7;13:95. doi: 10.3389/fnins.2019.00095. eCollection 2019.
6
CIFAR10-DVS: An Event-Stream Dataset for Object Classification.CIFAR10-DVS:用于目标分类的事件流数据集。
Front Neurosci. 2017 May 30;11:309. doi: 10.3389/fnins.2017.00309. eCollection 2017.
7
Training Deep Spiking Neural Networks Using Backpropagation.使用反向传播训练深度脉冲神经网络。
Front Neurosci. 2016 Nov 8;10:508. doi: 10.3389/fnins.2016.00508. eCollection 2016.
8
Converting Static Image Datasets to Spiking Neuromorphic Datasets Using Saccades.利用扫视将静态图像数据集转换为脉冲神经形态数据集
Front Neurosci. 2015 Nov 16;9:437. doi: 10.3389/fnins.2015.00437. eCollection 2015.
9
Is action potential threshold lowest in the axon?动作电位阈值在轴突中是最低的吗?
Nat Neurosci. 2008 Nov;11(11):1253-5. doi: 10.1038/nn.2203. Epub 2008 Oct 5.