• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

Spike-HAR++:一种用于基于事件的人类动作识别的高效节能轻量级并行脉冲变压器。

Spike-HAR++: an energy-efficient and lightweight parallel spiking transformer for event-based human action recognition.

作者信息

Lin Xinxu, Liu Mingxuan, Chen Hong

机构信息

School of Integrated Circuits, Tsinghua University, Beijing, China.

State Key Laboratory of Integrated Chips and Systems, Frontier Institute of Chip and System, Fudan University, Shanghai, China.

出版信息

Front Comput Neurosci. 2024 Nov 26;18:1508297. doi: 10.3389/fncom.2024.1508297. eCollection 2024.

DOI:10.3389/fncom.2024.1508297
PMID:39659428
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11628275/
Abstract

Event-based cameras are suitable for human action recognition (HAR) by providing movement perception with highly dynamic range, high temporal resolution, high power efficiency and low latency. Spike Neural Networks (SNNs) are naturally suited to deal with the asynchronous and sparse data from the event cameras due to their spike-based event-driven paradigm, with less power consumption compared to artificial neural networks. In this paper, we propose two end-to-end SNNs, namely Spike-HAR and Spike-HAR++, to introduce spiking transformer into event-based HAR. Spike-HAR includes two novel blocks: a spike attention branch, which enables model to focus on regions with high spike rates, reducing the impact of noise to improve the accuracy, and a parallel spike transformer block with simplified spiking self-attention mechanism, increasing computational efficiency. To better extract crucial information from high-level features, we modify the architecture of the spike attention branch and extend it in Spike-HAR to a higher dimension, proposing Spike-HAR++ to further enhance classification performance. Comprehensive experiments were conducted on four HAR datasets: SL-Animals-DVS, N-LSA64, DVS128 Gesture and DailyAction-DVS, to demonstrate the superior performance of our proposed model. Additionally, the proposed Spike-HAR and Spike-HAR++ require only 0.03 and 0.06 mJ, respectively, to process a sequence of event frames, with model sizes of only 0.7 and 1.8 M. This efficiency positions it as a promising new SNN baseline for the HAR community. Code is available at Spike-HAR++.

摘要

基于事件的相机通过提供具有高动态范围、高时间分辨率、高功率效率和低延迟的运动感知,适用于人类动作识别(HAR)。脉冲神经网络(SNN)由于其基于脉冲的事件驱动范式,天然适合处理来自事件相机的异步和稀疏数据,与人工神经网络相比功耗更低。在本文中,我们提出了两种端到端的SNN,即Spike-HAR和Spike-HAR++,将脉冲变换器引入基于事件的HAR中。Spike-HAR包括两个新颖的模块:一个脉冲注意力分支,使模型能够专注于高脉冲率区域,减少噪声影响以提高准确性;以及一个具有简化脉冲自注意力机制的并行脉冲变换器模块,提高计算效率。为了更好地从高级特征中提取关键信息,我们修改了脉冲注意力分支的架构,并在Spike-HAR中将其扩展到更高维度,提出了Spike-HAR++以进一步提高分类性能。我们在四个HAR数据集上进行了综合实验:SL-Animals-DVS、N-LSA64、DVS128 Gesture和DailyAction-DVS,以证明我们提出的模型的优越性能。此外,所提出的Spike-HAR和Spike-HAR++处理一系列事件帧分别仅需0.03和0.06 mJ,模型大小分别仅为0.7和1.8 M。这种效率使其成为HAR社区一个有前途的新SNN基线。代码可在Spike-HAR++获取。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5d74/11628275/ee171ace9380/fncom-18-1508297-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5d74/11628275/18199cfa77dd/fncom-18-1508297-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5d74/11628275/dcd827e54e8a/fncom-18-1508297-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5d74/11628275/35e0ab230dc7/fncom-18-1508297-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5d74/11628275/ee171ace9380/fncom-18-1508297-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5d74/11628275/18199cfa77dd/fncom-18-1508297-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5d74/11628275/dcd827e54e8a/fncom-18-1508297-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5d74/11628275/35e0ab230dc7/fncom-18-1508297-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5d74/11628275/ee171ace9380/fncom-18-1508297-g0004.jpg

相似文献

1
Spike-HAR++: an energy-efficient and lightweight parallel spiking transformer for event-based human action recognition.Spike-HAR++:一种用于基于事件的人类动作识别的高效节能轻量级并行脉冲变压器。
Front Comput Neurosci. 2024 Nov 26;18:1508297. doi: 10.3389/fncom.2024.1508297. eCollection 2024.
2
Sparser spiking activity can be better: Feature Refine-and-Mask spiking neural network for event-based visual recognition.稀疏尖峰活动可以更好:基于事件的视觉识别的特征细化和掩蔽尖峰神经网络。
Neural Netw. 2023 Sep;166:410-423. doi: 10.1016/j.neunet.2023.07.008. Epub 2023 Jul 20.
3
Spiking-PhysFormer: Camera-based remote photoplethysmography with parallel spike-driven transformer.Spiking-PhysFormer:基于摄像头的并行脉冲驱动变压器远程光电容积脉搏波描记法
Neural Netw. 2025 May;185:107128. doi: 10.1016/j.neunet.2025.107128. Epub 2025 Jan 10.
4
Optimizing Deeper Spiking Neural Networks for Dynamic Vision Sensing.深度尖峰神经网络在动态视觉传感中的优化。
Neural Netw. 2021 Dec;144:686-698. doi: 10.1016/j.neunet.2021.09.022. Epub 2021 Oct 5.
5
STSC-SNN: Spatio-Temporal Synaptic Connection with temporal convolution and attention for spiking neural networks.STSC-SNN:用于脉冲神经网络的具有时间卷积和注意力机制的时空突触连接
Front Neurosci. 2022 Dec 23;16:1079357. doi: 10.3389/fnins.2022.1079357. eCollection 2022.
6
SGLFormer: Spiking Global-Local-Fusion Transformer with high performance.SGLFormer:具有高性能的脉冲全局-局部融合变压器。
Front Neurosci. 2024 Mar 12;18:1371290. doi: 10.3389/fnins.2024.1371290. eCollection 2024.
7
Boost event-driven tactile learning with location spiking neurons.利用位置发放神经元增强事件驱动的触觉学习。
Front Neurosci. 2023 Apr 21;17:1127537. doi: 10.3389/fnins.2023.1127537. eCollection 2023.
8
[A bio-inspired hierarchical spiking neural network with biological synaptic plasticity for event camera object recognition].一种具有生物突触可塑性的用于事件相机目标识别的生物启发式分层脉冲神经网络
Sheng Wu Yi Xue Gong Cheng Xue Za Zhi. 2023 Aug 25;40(4):692-699. doi: 10.7507/1001-5515.202207040.
9
Efficient human activity recognition with spatio-temporal spiking neural networks.基于时空脉冲神经网络的高效人类活动识别
Front Neurosci. 2023 Sep 14;17:1233037. doi: 10.3389/fnins.2023.1233037. eCollection 2023.
10
Heterogeneous recurrent spiking neural network for spatio-temporal classification.用于时空分类的异构递归脉冲神经网络
Front Neurosci. 2023 Jan 30;17:994517. doi: 10.3389/fnins.2023.994517. eCollection 2023.

本文引用的文献

1
Advancing Spiking Neural Networks Toward Deep Residual Learning.推动脉冲神经网络迈向深度残差学习
IEEE Trans Neural Netw Learn Syst. 2025 Feb;36(2):2353-2367. doi: 10.1109/TNNLS.2024.3355393. Epub 2025 Feb 6.
2
SpikingJelly: An open-source machine learning infrastructure platform for spike-based intelligence.SpikingJelly:一个用于基于尖峰的智能的开源机器学习基础架构平台。
Sci Adv. 2023 Oct 6;9(40):eadi1480. doi: 10.1126/sciadv.adi1480.
3
Event Transformer . A Multi-Purpose Solution for Efficient Event Data Processing.事件变换器:一种用于高效事件数据处理的多功能解决方案
IEEE Trans Pattern Anal Mach Intell. 2023 Dec;45(12):16013-16020. doi: 10.1109/TPAMI.2023.3311336. Epub 2023 Nov 3.
4
Action Recognition and Benchmark Using Event Cameras.使用事件相机的动作识别与基准测试
IEEE Trans Pattern Anal Mach Intell. 2023 Dec;45(12):14081-14097. doi: 10.1109/TPAMI.2023.3300741. Epub 2023 Nov 3.
5
A Spatial-Channel-Temporal-Fused Attention for Spiking Neural Networks.一种用于尖峰神经网络的空间-通道-时间融合注意力机制。
IEEE Trans Neural Netw Learn Syst. 2024 Oct;35(10):14315-14329. doi: 10.1109/TNNLS.2023.3278265. Epub 2024 Oct 7.
6
Human Action Recognition From Various Data Modalities: A Review.基于多种数据模态的人类行为识别综述
IEEE Trans Pattern Anal Mach Intell. 2023 Mar;45(3):3200-3225. doi: 10.1109/TPAMI.2022.3183112. Epub 2023 Feb 3.
7
Time-Ordered Recent Event (TORE) Volumes for Event Cameras.事件相机的时间顺序近期事件(TORE)体积
IEEE Trans Pattern Anal Mach Intell. 2023 Feb;45(2):2519-2532. doi: 10.1109/TPAMI.2022.3172212. Epub 2023 Jan 6.
8
A Survey on Vision Transformer.视觉Transformer综述
IEEE Trans Pattern Anal Mach Intell. 2023 Jan;45(1):87-110. doi: 10.1109/TPAMI.2022.3152247. Epub 2022 Dec 5.
9
Spiking Deep Residual Networks.尖峰深度残差网络
IEEE Trans Neural Netw Learn Syst. 2023 Aug;34(8):5200-5205. doi: 10.1109/TNNLS.2021.3119238. Epub 2023 Aug 4.
10
Event-Stream Representation for Human Gaits Identification Using Deep Neural Networks.基于深度神经网络的人类步态识别的事件流表示。
IEEE Trans Pattern Anal Mach Intell. 2022 Jul;44(7):3436-3449. doi: 10.1109/TPAMI.2021.3054886. Epub 2022 Jun 3.