• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

具有短期突触可塑性的尖峰神经元形成了优越的生成性网络。

Spiking neurons with short-term synaptic plasticity form superior generative networks.

机构信息

Kirchhoff Institute for Physics, University of Heidelberg, Heidelberg, Germany.

Department of Physiology, University of Bern, Bern, Switzerland.

出版信息

Sci Rep. 2018 Jul 13;8(1):10651. doi: 10.1038/s41598-018-28999-2.

DOI:10.1038/s41598-018-28999-2
PMID:30006554
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC6045624/
Abstract

Spiking networks that perform probabilistic inference have been proposed both as models of cortical computation and as candidates for solving problems in machine learning. However, the evidence for spike-based computation being in any way superior to non-spiking alternatives remains scarce. We propose that short-term synaptic plasticity can provide spiking networks with distinct computational advantages compared to their classical counterparts. When learning from high-dimensional, diverse datasets, deep attractors in the energy landscape often cause mixing problems to the sampling process. Classical algorithms solve this problem by employing various tempering techniques, which are both computationally demanding and require global state updates. We demonstrate how similar results can be achieved in spiking networks endowed with local short-term synaptic plasticity. Additionally, we discuss how these networks can even outperform tempering-based approaches when the training data is imbalanced. We thereby uncover a powerful computational property of the biologically inspired, local, spike-triggered synaptic dynamics based simply on a limited pool of synaptic resources, which enables them to deal with complex sensory data.

摘要

提出了执行概率推理的尖峰网络,既是皮质计算的模型,也是机器学习中解决问题的候选方案。然而,基于尖峰的计算在任何方面都优于非尖峰替代方案的证据仍然很少。我们提出,短期突触可塑性可以为尖峰网络提供与经典网络相比具有明显计算优势的方法。在从高维、多样化的数据集学习时,能量景观中的深度吸引子经常导致采样过程中的混合问题。经典算法通过采用各种回火技术来解决这个问题,这些技术既计算量大,又需要全局状态更新。我们展示了如何在具有局部短期突触可塑性的尖峰网络中实现类似的结果。此外,当训练数据不平衡时,我们还讨论了这些网络如何甚至可以超过基于回火的方法。因此,我们揭示了一种基于生物启发的、局部的、基于尖峰触发的突触动力学的强大计算特性,这种特性基于有限的突触资源池,使它们能够处理复杂的感官数据。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/599c/6045624/0c2091fdec9c/41598_2018_28999_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/599c/6045624/8161e8219827/41598_2018_28999_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/599c/6045624/810f6c171049/41598_2018_28999_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/599c/6045624/09bc29245769/41598_2018_28999_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/599c/6045624/1ce86cd3d0a2/41598_2018_28999_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/599c/6045624/0c2091fdec9c/41598_2018_28999_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/599c/6045624/8161e8219827/41598_2018_28999_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/599c/6045624/810f6c171049/41598_2018_28999_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/599c/6045624/09bc29245769/41598_2018_28999_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/599c/6045624/1ce86cd3d0a2/41598_2018_28999_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/599c/6045624/0c2091fdec9c/41598_2018_28999_Fig5_HTML.jpg

相似文献

1
Spiking neurons with short-term synaptic plasticity form superior generative networks.具有短期突触可塑性的尖峰神经元形成了优越的生成性网络。
Sci Rep. 2018 Jul 13;8(1):10651. doi: 10.1038/s41598-018-28999-2.
2
Synaptic dynamics: linear model and adaptation algorithm.突触动力学:线性模型与自适应算法。
Neural Netw. 2014 Aug;56:49-68. doi: 10.1016/j.neunet.2014.04.001. Epub 2014 Apr 28.
3
Meta-SpikePropamine: learning to learn with synaptic plasticity in spiking neural networks.元尖峰传播胺:在脉冲神经网络中通过突触可塑性进行学习学习
Front Neurosci. 2023 May 12;17:1183321. doi: 10.3389/fnins.2023.1183321. eCollection 2023.
4
Stochasticity from function - Why the Bayesian brain may need no noise.从功能角度看随机性——为什么贝叶斯大脑可能不需要噪声。
Neural Netw. 2019 Nov;119:200-213. doi: 10.1016/j.neunet.2019.08.002. Epub 2019 Aug 19.
5
Mirrored STDP Implements Autoencoder Learning in a Network of Spiking Neurons.镜像脉冲时间依赖可塑性在脉冲神经元网络中实现自动编码器学习。
PLoS Comput Biol. 2015 Dec 3;11(12):e1004566. doi: 10.1371/journal.pcbi.1004566. eCollection 2015 Dec.
6
Solving Constraint Satisfaction Problems with Networks of Spiking Neurons.利用脉冲神经元网络解决约束满足问题。
Front Neurosci. 2016 Mar 30;10:118. doi: 10.3389/fnins.2016.00118. eCollection 2016.
7
Hierarchical Bayesian Inference and Learning in Spiking Neural Networks.分层贝叶斯推断和尖峰神经网络中的学习。
IEEE Trans Cybern. 2019 Jan;49(1):133-145. doi: 10.1109/TCYB.2017.2768554. Epub 2017 Nov 9.
8
Stochastic variational learning in recurrent spiking networks.递归尖峰网络中的随机变分学习。
Front Comput Neurosci. 2014 Apr 4;8:38. doi: 10.3389/fncom.2014.00038. eCollection 2014.
9
Efficient computation based on stochastic spikes.基于随机脉冲的高效计算。
Neural Comput. 2007 May;19(5):1313-43. doi: 10.1162/neco.2007.19.5.1313.
10
Stochastic Synapses Enable Efficient Brain-Inspired Learning Machines.随机突触助力高效受脑启发的学习机器。
Front Neurosci. 2016 Jun 29;10:241. doi: 10.3389/fnins.2016.00241. eCollection 2016.

引用本文的文献

1
CMOS Single-Photon Avalanche Diode Circuits for Probabilistic Computing.用于概率计算的互补金属氧化物半导体单光子雪崩二极管电路
IEEE J Explor Solid State Comput Devices Circuits. 2024;10:49-57. doi: 10.1109/jxcdc.2024.3452030. Epub 2024 Aug 29.
2
Advanced Neuromorphic Applications Enabled by Synaptic Ion-Gating Vertical Transistors.突触离子门控垂直晶体管实现的先进神经形态应用
Adv Sci (Weinh). 2024 Jul;11(27):e2305611. doi: 10.1002/advs.202305611. Epub 2024 May 17.
3
A surrogate gradient spiking baseline for speech command recognition.

本文引用的文献

1
Event-Driven Random Back-Propagation: Enabling Neuromorphic Deep Learning Machines.事件驱动的随机反向传播:助力神经形态深度学习机器
Front Neurosci. 2017 Jun 21;11:324. doi: 10.3389/fnins.2017.00324. eCollection 2017.
2
Deep Neural Networks: A New Framework for Modeling Biological Vision and Brain Information Processing.深度神经网络:一种用于模拟生物视觉和大脑信息处理的新框架。
Annu Rev Vis Sci. 2015 Nov 24;1:417-446. doi: 10.1146/annurev-vision-082114-035447.
3
Training Deep Spiking Neural Networks Using Backpropagation.使用反向传播训练深度脉冲神经网络。
用于语音命令识别的替代梯度脉冲基线
Front Neurosci. 2022 Aug 22;16:865897. doi: 10.3389/fnins.2022.865897. eCollection 2022.
4
Variational learning of quantum ground states on spiking neuromorphic hardware.基于脉冲神经形态硬件的量子基态变分学习。
iScience. 2022 Jul 5;25(8):104707. doi: 10.1016/j.isci.2022.104707. eCollection 2022 Aug 19.
5
Cortical oscillations support sampling-based computations in spiking neural networks.皮层振荡支持基于抽样的尖峰神经网络计算。
PLoS Comput Biol. 2022 Mar 24;18(3):e1009753. doi: 10.1371/journal.pcbi.1009753. eCollection 2022 Mar.
6
Spike-induced ordering: Stochastic neural spikes provide immediate adaptability to the sensorimotor system.尖峰诱导有序:随机神经尖峰为感觉运动系统提供即时的适应性。
Proc Natl Acad Sci U S A. 2020 Jun 2;117(22):12486-12496. doi: 10.1073/pnas.1819707117. Epub 2020 May 19.
7
Accelerated Physical Emulation of Bayesian Inference in Spiking Neural Networks.脉冲神经网络中贝叶斯推理的加速物理仿真
Front Neurosci. 2019 Nov 14;13:1201. doi: 10.3389/fnins.2019.01201. eCollection 2019.
Front Neurosci. 2016 Nov 8;10:508. doi: 10.3389/fnins.2016.00508. eCollection 2016.
4
Stochastic inference with spiking neurons in the high-conductance state.高电导状态下脉冲神经元的随机推理。
Phys Rev E. 2016 Oct;94(4-1):042312. doi: 10.1103/PhysRevE.94.042312. Epub 2016 Oct 20.
5
Random synaptic feedback weights support error backpropagation for deep learning.随机突触反馈权重支持深度学习的误差反向传播。
Nat Commun. 2016 Nov 8;7:13276. doi: 10.1038/ncomms13276.
6
Linking pattern completion in the hippocampus to predictive coding in visual cortex.将海马体中的模式完成与视觉皮层中的预测编码联系起来。
Nat Neurosci. 2016 May;19(5):665-667. doi: 10.1038/nn.4284. Epub 2016 Apr 11.
7
Deep learning.深度学习。
Nature. 2015 May 28;521(7553):436-44. doi: 10.1038/nature14539.
8
Deep learning in neural networks: an overview.神经网络中的深度学习:综述。
Neural Netw. 2015 Jan;61:85-117. doi: 10.1016/j.neunet.2014.09.003. Epub 2014 Oct 13.
9
Six networks on a universal neuromorphic computing substrate.六个网络在一个通用的神经形态计算基板上。
Front Neurosci. 2013 Feb 18;7:11. doi: 10.3389/fnins.2013.00011. eCollection 2013.
10
Neural dynamics as sampling: a model for stochastic computation in recurrent networks of spiking neurons.神经动力学作为采样:脉冲神经元的递归网络中随机计算的模型。
PLoS Comput Biol. 2011 Nov;7(11):e1002211. doi: 10.1371/journal.pcbi.1002211. Epub 2011 Nov 3.