• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

DL-ReSuMe:一种基于延迟学习的尖峰神经元远程监督方法。

DL-ReSuMe: A Delay Learning-Based Remote Supervised Method for Spiking Neurons.

出版信息

IEEE Trans Neural Netw Learn Syst. 2015 Dec;26(12):3137-49. doi: 10.1109/TNNLS.2015.2404938. Epub 2015 Mar 18.

DOI:10.1109/TNNLS.2015.2404938
PMID:25794401
Abstract

Recent research has shown the potential capability of spiking neural networks (SNNs) to model complex information processing in the brain. There is biological evidence to prove the use of the precise timing of spikes for information coding. However, the exact learning mechanism in which the neuron is trained to fire at precise times remains an open problem. The majority of the existing learning methods for SNNs are based on weight adjustment. However, there is also biological evidence that the synaptic delay is not constant. In this paper, a learning method for spiking neurons, called delay learning remote supervised method (DL-ReSuMe), is proposed to merge the delay shift approach and ReSuMe-based weight adjustment to enhance the learning performance. DL-ReSuMe uses more biologically plausible properties, such as delay learning, and needs less weight adjustment than ReSuMe. Simulation results have shown that the proposed DL-ReSuMe approach achieves learning accuracy and learning speed improvements compared with ReSuMe.

摘要

最近的研究表明,尖峰神经网络(SNN)具有模拟大脑中复杂信息处理的潜力。有生物学证据证明,利用尖峰的精确时间进行信息编码。然而,神经元被训练在精确时间发射的确切学习机制仍然是一个悬而未决的问题。大多数现有的 SNN 学习方法都是基于权重调整的。然而,也有生物学证据表明突触延迟不是恒定的。在本文中,提出了一种用于尖峰神经元的学习方法,称为延迟学习远程监督方法(DL-ReSuMe),该方法将延迟移位方法和基于 ReSuMe 的权重调整相结合,以提高学习性能。DL-ReSuMe 使用了更具生物学合理性的特性,如延迟学习,并且比 ReSuMe 需要更少的权重调整。仿真结果表明,与 ReSuMe 相比,所提出的 DL-ReSuMe 方法在学习准确性和学习速度方面都有所提高。

相似文献

1
DL-ReSuMe: A Delay Learning-Based Remote Supervised Method for Spiking Neurons.DL-ReSuMe:一种基于延迟学习的尖峰神经元远程监督方法。
IEEE Trans Neural Netw Learn Syst. 2015 Dec;26(12):3137-49. doi: 10.1109/TNNLS.2015.2404938. Epub 2015 Mar 18.
2
Supervised learning in spiking neural networks with ReSuMe: sequence learning, classification, and spike shifting.基于 ReSuMe 的尖峰神经网络监督学习:序列学习、分类和尖峰转移。
Neural Comput. 2010 Feb;22(2):467-510. doi: 10.1162/neco.2009.11-08-901.
3
A Cross-Correlated Delay Shift Supervised Learning Method for Spiking Neurons with Application to Interictal Spike Detection in Epilepsy.一种用于脉冲神经元的互相关延迟移位监督学习方法及其在癫痫发作间期棘波检测中的应用
Int J Neural Syst. 2017 May;27(3):1750002. doi: 10.1142/S0129065717500022. Epub 2016 Sep 1.
4
A supervised multi-spike learning algorithm based on gradient descent for spiking neural networks.基于梯度下降的监督多尖峰学习算法在尖峰神经网络中的应用。
Neural Netw. 2013 Jul;43:99-113. doi: 10.1016/j.neunet.2013.02.003. Epub 2013 Feb 16.
5
Spiking neural networks.脉冲神经网络。
Int J Neural Syst. 2009 Aug;19(4):295-308. doi: 10.1142/S0129065709002002.
6
A new supervised learning algorithm for spiking neurons.一种新的用于尖峰神经元的监督学习算法。
Neural Comput. 2013 Jun;25(6):1472-511. doi: 10.1162/NECO_a_00450. Epub 2013 Mar 21.
7
Synaptic dynamics: linear model and adaptation algorithm.突触动力学:线性模型与自适应算法。
Neural Netw. 2014 Aug;56:49-68. doi: 10.1016/j.neunet.2014.04.001. Epub 2014 Apr 28.
8
An STDP training algorithm for a spiking neural network with dynamic threshold neurons.一种具有动态阈值神经元的尖峰神经网络的 STDP 训练算法。
Int J Neural Syst. 2010 Dec;20(6):463-80. doi: 10.1142/S0129065710002553.
9
An online supervised learning method based on gradient descent for spiking neurons.一种基于梯度下降的用于脉冲神经元的在线监督学习方法。
Neural Netw. 2017 Sep;93:7-20. doi: 10.1016/j.neunet.2017.04.010. Epub 2017 Apr 27.
10
Span: spike pattern association neuron for learning spatio-temporal spike patterns.用于学习时空尖峰模式的尖峰模式关联神经元。
Int J Neural Syst. 2012 Aug;22(4):1250012. doi: 10.1142/S0129065712500128. Epub 2012 Jul 12.

引用本文的文献

1
Brain-inspired learning rules for spiking neural network-based control: a tutorial.基于脉冲神经网络控制的受脑启发学习规则:教程
Biomed Eng Lett. 2024 Dec 2;15(1):37-55. doi: 10.1007/s13534-024-00436-6. eCollection 2025 Jan.
2
Learning delays through gradients and structure: emergence of spatiotemporal patterns in spiking neural networks.通过梯度和结构实现学习延迟:脉冲神经网络中时空模式的出现。
Front Comput Neurosci. 2024 Dec 20;18:1460309. doi: 10.3389/fncom.2024.1460309. eCollection 2024.
3
Training multi-layer spiking neural networks with plastic synaptic weights and delays.
使用可塑性突触权重和延迟训练多层脉冲神经网络。
Front Neurosci. 2024 Jan 24;17:1253830. doi: 10.3389/fnins.2023.1253830. eCollection 2023.
4
Learnable axonal delay in spiking neural networks improves spoken word recognition.脉冲神经网络中可学习的轴突延迟改善了口语单词识别。
Front Neurosci. 2023 Nov 9;17:1275944. doi: 10.3389/fnins.2023.1275944. eCollection 2023.
5
-A Spiking Neuron-Based Classifier That Combines Weight-Adjustment and Delay-Shift.一种结合权重调整和延迟移位的基于脉冲神经元的分类器。
Front Neurosci. 2021 Feb 19;15:582608. doi: 10.3389/fnins.2021.582608. eCollection 2021.
6
SpiFoG: an efficient supervised learning algorithm for the network of spiking neurons.SpiFoG:一种用于尖峰神经元网络的高效监督学习算法。
Sci Rep. 2020 Aug 4;10(1):13122. doi: 10.1038/s41598-020-70136-5.
7
Event-Based Computation for Touch Localization Based on Precise Spike Timing.基于精确脉冲时间的触摸定位的基于事件的计算
Front Neurosci. 2020 May 19;14:420. doi: 10.3389/fnins.2020.00420. eCollection 2020.
8
An Efficient and Perceptually Motivated Auditory Neural Encoding and Decoding Algorithm for Spiking Neural Networks.一种用于脉冲神经网络的高效且基于感知动机的听觉神经编码与解码算法。
Front Neurosci. 2020 Jan 22;13:1420. doi: 10.3389/fnins.2019.01420. eCollection 2019.
9
First Error-Based Supervised Learning Algorithm for Spiking Neural Networks.用于脉冲神经网络的首个基于误差的监督学习算法。
Front Neurosci. 2019 Jun 6;13:559. doi: 10.3389/fnins.2019.00559. eCollection 2019.
10
A Delay Learning Algorithm Based on Spike Train Kernels for Spiking Neurons.一种基于脉冲序列核的脉冲神经元延迟学习算法。
Front Neurosci. 2019 Mar 27;13:252. doi: 10.3389/fnins.2019.00252. eCollection 2019.