• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于能量模型中反向传播的STDP兼容近似

STDP-Compatible Approximation of Backpropagation in an Energy-Based Model.

作者信息

Bengio Yoshua, Mesnard Thomas, Fischer Asja, Zhang Saizheng, Wu Yuhuai

机构信息

Montreal Institute for Learning Algorithms, University of Montreal, Montreal H3T 1J4, Quebec, Canada, and Canadian Institute for Advanced Research

Computer Science Department, École Normale Supérieure, Paris 75005, France

出版信息

Neural Comput. 2017 Mar;29(3):555-577. doi: 10.1162/NECO_a_00934. Epub 2017 Jan 17.

DOI:10.1162/NECO_a_00934
PMID:28095200
Abstract

We show that Langevin Markov chain Monte Carlo inference in an energy-based model with latent variables has the property that the early steps of inference, starting from a stationary point, correspond to propagating error gradients into internal layers, similar to backpropagation. The backpropagated error is with respect to output units that have received an outside driving force pushing them away from the stationary point. Backpropagated error gradients correspond to temporal derivatives with respect to the activation of hidden units. These lead to a weight update proportional to the product of the presynaptic firing rate and the temporal rate of change of the postsynaptic firing rate. Simulations and a theoretical argument suggest that this rate-based update rule is consistent with those associated with spike-timing-dependent plasticity. The ideas presented in this article could be an element of a theory for explaining how brains perform credit assignment in deep hierarchies as efficiently as backpropagation does, with neural computation corresponding to both approximate inference in continuous-valued latent variables and error backpropagation, at the same time.

摘要

我们表明,在具有潜在变量的基于能量的模型中,朗之万马尔可夫链蒙特卡罗推理具有这样的特性:从一个稳定点开始的推理早期步骤,类似于反向传播,对应于将误差梯度传播到内部层。反向传播的误差是相对于那些接收到外部驱动力从而使其远离稳定点的输出单元而言的。反向传播的误差梯度对应于关于隐藏单元激活的时间导数。这些导致权重更新与突触前发放率和突触后发放率的时间变化率的乘积成比例。模拟和理论论证表明,这种基于速率的更新规则与那些与尖峰时间依赖可塑性相关的规则是一致的。本文提出的观点可能是一种理论的要素,该理论用于解释大脑如何像反向传播一样高效地在深度层次结构中进行信用分配,其中神经计算同时对应于连续值潜在变量中的近似推理和误差反向传播。

相似文献

1
STDP-Compatible Approximation of Backpropagation in an Energy-Based Model.基于能量模型中反向传播的STDP兼容近似
Neural Comput. 2017 Mar;29(3):555-577. doi: 10.1162/NECO_a_00934. Epub 2017 Jan 17.
2
Equilibrium Propagation: Bridging the Gap between Energy-Based Models and Backpropagation.平衡传播:弥合基于能量模型与反向传播之间的差距
Front Comput Neurosci. 2017 May 4;11:24. doi: 10.3389/fncom.2017.00024. eCollection 2017.
3
A learning theory for reward-modulated spike-timing-dependent plasticity with application to biofeedback.一种用于奖励调制的依赖于尖峰时间的可塑性的学习理论及其在生物反馈中的应用。
PLoS Comput Biol. 2008 Oct;4(10):e1000180. doi: 10.1371/journal.pcbi.1000180. Epub 2008 Oct 10.
4
Learning cortical hierarchies with temporal Hebbian updates.通过时间赫布更新学习皮层层次结构。
Front Comput Neurosci. 2023 May 24;17:1136010. doi: 10.3389/fncom.2023.1136010. eCollection 2023.
5
How the shape of pre- and postsynaptic signals can influence STDP: a biophysical model.突触前和突触后信号的形状如何影响突触可塑性:一个生物物理模型。
Neural Comput. 2004 Mar;16(3):595-625. doi: 10.1162/089976604772744929.
6
Modulation of synaptic plasticity by the coactivation of spatially distinct synaptic inputs in rat hippocampal CA1 apical dendrites.大鼠海马 CA1 树突锥体上支不同空间部位突触输入的共激活对突触可塑性的调制。
Brain Res. 2013 Aug 14;1526:1-14. doi: 10.1016/j.brainres.2013.05.023. Epub 2013 May 24.
7
Reconciling the STDP and BCM models of synaptic plasticity in a spiking recurrent neural network.在一个尖峰循环神经网络中协调 STDP 和 BCM 模型的突触可塑性。
Neural Comput. 2010 Aug;22(8):2059-85. doi: 10.1162/NECO_a_00003-Bush.
8
Reinforcement learning through modulation of spike-timing-dependent synaptic plasticity.通过调节尖峰时间依赖性突触可塑性进行强化学习。
Neural Comput. 2007 Jun;19(6):1468-502. doi: 10.1162/neco.2007.19.6.1468.
9
Triplet Spike Time-Dependent Plasticity in a Floating-Gate Synapse.浮栅突触中的三重尖峰时间依赖性可塑性
IEEE Trans Neural Netw Learn Syst. 2015 Dec 24;28(4). doi: 10.1109/TNNLS.2015.2506740. Print 2017 Apr.
10
A theory of cerebellar cortex and adaptive motor control based on two types of universal function approximation capability.基于两种通用函数逼近能力的小脑皮质与适应性运动控制理论。
Neural Netw. 2016 Mar;75:173-96. doi: 10.1016/j.neunet.2015.12.012. Epub 2016 Jan 6.

引用本文的文献

1
Research on target detection for autonomous driving based on ECS-spiking neural networks.基于ECS脉冲神经网络的自动驾驶目标检测研究
Sci Rep. 2025 Apr 21;15(1):13725. doi: 10.1038/s41598-025-97913-4.
2
Temporal Contrastive Learning through implicit non-equilibrium memory.通过隐式非平衡记忆实现时间对比学习。
Nat Commun. 2025 Mar 4;16(1):2163. doi: 10.1038/s41467-025-57043-x.
3
Free-space optical spiking neural network.自由空间光学脉冲神经网络。
PLoS One. 2024 Dec 30;19(12):e0313547. doi: 10.1371/journal.pone.0313547. eCollection 2024.
4
An Agent-Based Model to Reproduce the Boolean Logic Behaviour of Neuronal Self-Organised Communities through Pulse Delay Modulation and Generation of Logic Gates.一种基于代理的模型,通过脉冲延迟调制和逻辑门生成来再现神经元自组织群落的布尔逻辑行为。
Biomimetics (Basel). 2024 Feb 9;9(2):101. doi: 10.3390/biomimetics9020101.
5
A brain-inspired algorithm that mitigates catastrophic forgetting of artificial and spiking neural networks with low computational cost.一种具有低计算成本的脑启发算法,可减轻人工和尖峰神经网络的灾难性遗忘。
Sci Adv. 2023 Aug 25;9(34):eadi2947. doi: 10.1126/sciadv.adi2947.
6
Learning cortical hierarchies with temporal Hebbian updates.通过时间赫布更新学习皮层层次结构。
Front Comput Neurosci. 2023 May 24;17:1136010. doi: 10.3389/fncom.2023.1136010. eCollection 2023.
7
Convergence of Artificial Intelligence and Neuroscience towards the Diagnosis of Neurological Disorders-A Scoping Review.人工智能与神经科学在神经紊乱诊断中的交汇:综述
Sensors (Basel). 2023 Mar 13;23(6):3062. doi: 10.3390/s23063062.
8
Supervised Learning Algorithm Based on Spike Train Inner Product for Deep Spiking Neural Networks.基于脉冲序列内积的深度脉冲神经网络监督学习算法
Brain Sci. 2023 Jan 18;13(2):168. doi: 10.3390/brainsci13020168.
9
Training spiking neuronal networks to perform motor control using reinforcement and evolutionary learning.利用强化学习和进化学习训练脉冲神经网络以执行运动控制。
Front Comput Neurosci. 2022 Sep 30;16:1017284. doi: 10.3389/fncom.2022.1017284. eCollection 2022.
10
Neurons learn by predicting future activity.神经元通过预测未来的活动来学习。
Nat Mach Intell. 2022 Jan;4(1):62-72. doi: 10.1038/s42256-021-00430-y. Epub 2022 Jan 25.