• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

平衡传播:弥合基于能量模型与反向传播之间的差距

Equilibrium Propagation: Bridging the Gap between Energy-Based Models and Backpropagation.

作者信息

Scellier Benjamin, Bengio Yoshua

机构信息

Département d'Informatique et de Recherche Opérationnelle, Montreal Institute for Learning Algorithms, Université de MontréalMontreal, QC, Canada.

出版信息

Front Comput Neurosci. 2017 May 4;11:24. doi: 10.3389/fncom.2017.00024. eCollection 2017.

DOI:10.3389/fncom.2017.00024
PMID:28522969
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC5415673/
Abstract

We introduce Equilibrium Propagation, a learning framework for energy-based models. It involves only one kind of neural computation, performed in both the first phase (when the prediction is made) and the second phase of training (after the target or prediction error is revealed). Although this algorithm computes the gradient of an objective function just like Backpropagation, it does not need a special computation or circuit for the second phase, where errors are implicitly propagated. Equilibrium Propagation shares similarities with Contrastive Hebbian Learning and Contrastive Divergence while solving the theoretical issues of both algorithms: our algorithm computes the gradient of a well-defined objective function. Because the objective function is defined in terms of local perturbations, the second phase of Equilibrium Propagation corresponds to only nudging the prediction (fixed point or stationary distribution) toward a configuration that reduces prediction error. In the case of a recurrent multi-layer supervised network, the output units are slightly nudged toward their target in the second phase, and the perturbation introduced at the output layer propagates backward in the hidden layers. We show that the signal "back-propagated" during this second phase corresponds to the propagation of error derivatives and encodes the gradient of the objective function, when the synaptic update corresponds to a standard form of spike-timing dependent plasticity. This work makes it more plausible that a mechanism similar to Backpropagation could be implemented by brains, since leaky integrator neural computation performs both inference and error back-propagation in our model. The only local difference between the two phases is whether synaptic changes are allowed or not. We also show experimentally that multi-layer recurrently connected networks with 1, 2, and 3 hidden layers can be trained by Equilibrium Propagation on the permutation-invariant MNIST task.

摘要

我们引入了平衡传播,这是一种基于能量模型的学习框架。它只涉及一种神经计算,在训练的第一阶段(进行预测时)和第二阶段(目标或预测误差被揭示后)都要执行。尽管此算法像反向传播一样计算目标函数的梯度,但在第二阶段不需要特殊的计算或电路,误差在该阶段是隐式传播的。平衡传播与对比赫布学习和对比散度有相似之处,同时解决了这两种算法的理论问题:我们的算法计算一个定义明确的目标函数的梯度。由于目标函数是根据局部扰动定义的,平衡传播的第二阶段仅对应于将预测(不动点或平稳分布)朝着减少预测误差的配置微调。在递归多层监督网络的情况下,输出单元在第二阶段会朝着其目标轻微微调,并且在输出层引入的扰动会在隐藏层中向后传播。我们表明,当突触更新对应于标准形式的脉冲时间依赖可塑性时,在第二阶段“反向传播”的信号对应于误差导数的传播,并编码目标函数的梯度。这项工作使得大脑有可能实现类似于反向传播的机制,因为在我们的模型中,泄漏积分器神经计算同时执行推理和误差反向传播。两个阶段之间唯一的局部差异在于是否允许突触变化。我们还通过实验表明,可以使用平衡传播在排列不变的MNIST任务上训练具有1个、2个和3个隐藏层的多层递归连接网络。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2287/5415673/a4b57ff1f73c/fncom-11-00024-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2287/5415673/0051c08aa421/fncom-11-00024-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2287/5415673/0e565c8c8266/fncom-11-00024-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2287/5415673/a4b57ff1f73c/fncom-11-00024-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2287/5415673/0051c08aa421/fncom-11-00024-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2287/5415673/0e565c8c8266/fncom-11-00024-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2287/5415673/a4b57ff1f73c/fncom-11-00024-g0003.jpg

相似文献

1
Equilibrium Propagation: Bridging the Gap between Energy-Based Models and Backpropagation.平衡传播:弥合基于能量模型与反向传播之间的差距
Front Comput Neurosci. 2017 May 4;11:24. doi: 10.3389/fncom.2017.00024. eCollection 2017.
2
Biologically Plausible Training Mechanisms for Self-Supervised Learning in Deep Networks.深度网络中自监督学习的生物学合理训练机制
Front Comput Neurosci. 2022 Mar 21;16:789253. doi: 10.3389/fncom.2022.789253. eCollection 2022.
3
Scaling Equilibrium Propagation to Deep ConvNets by Drastically Reducing Its Gradient Estimator Bias.通过大幅降低梯度估计偏差将平衡传播扩展到深度卷积神经网络
Front Neurosci. 2021 Feb 18;15:633674. doi: 10.3389/fnins.2021.633674. eCollection 2021.
4
Equivalence of Equilibrium Propagation and Recurrent Backpropagation.平衡传播与循环反向传播的等效性。
Neural Comput. 2019 Feb;31(2):312-329. doi: 10.1162/neco_a_01160. Epub 2018 Dec 21.
5
Biologically plausible deep learning - But how far can we go with shallow networks?生物学上合理的深度学习——但我们可以在浅层网络中走多远?
Neural Netw. 2019 Oct;118:90-101. doi: 10.1016/j.neunet.2019.06.001. Epub 2019 Jun 20.
6
Equilibrium Propagation for Memristor-Based Recurrent Neural Networks.基于忆阻器的递归神经网络的平衡传播
Front Neurosci. 2020 Mar 24;14:240. doi: 10.3389/fnins.2020.00240. eCollection 2020.
7
Equivalence of backpropagation and contrastive Hebbian learning in a layered network.分层网络中反向传播与对比赫布学习的等效性。
Neural Comput. 2003 Feb;15(2):441-54. doi: 10.1162/089976603762552988.
8
Tuning Convolutional Spiking Neural Network With Biologically Plausible Reward Propagation.基于生物合理奖励传播的卷积脉冲神经网络调优
IEEE Trans Neural Netw Learn Syst. 2022 Dec;33(12):7621-7631. doi: 10.1109/TNNLS.2021.3085966. Epub 2022 Nov 30.
9
Learning cortical hierarchies with temporal Hebbian updates.通过时间赫布更新学习皮层层次结构。
Front Comput Neurosci. 2023 May 24;17:1136010. doi: 10.3389/fncom.2023.1136010. eCollection 2023.
10
Contrastive Hebbian learning with random feedback weights.对比随机反馈权重的Hebbian 学习。
Neural Netw. 2019 Jun;114:1-14. doi: 10.1016/j.neunet.2019.01.008. Epub 2019 Feb 21.

引用本文的文献

1
Effective methods and framework for energy-based local learning of deep neural networks.基于能量的深度神经网络局部学习的有效方法与框架。
Front Artif Intell. 2025 Aug 26;8:1605706. doi: 10.3389/frai.2025.1605706. eCollection 2025.
2
Training of physical neural networks.物理神经网络的训练。
Nature. 2025 Sep;645(8079):53-61. doi: 10.1038/s41586-025-09384-2. Epub 2025 Sep 3.
3
Reward-optimizing learning using stochastic release plasticity.利用随机释放可塑性进行奖励优化学习。

本文引用的文献

1
STDP-Compatible Approximation of Backpropagation in an Energy-Based Model.基于能量模型中反向传播的STDP兼容近似
Neural Comput. 2017 Mar;29(3):555-577. doi: 10.1162/NECO_a_00934. Epub 2017 Jan 17.
2
Random synaptic feedback weights support error backpropagation for deep learning.随机突触反馈权重支持深度学习的误差反向传播。
Nat Commun. 2016 Nov 8;7:13276. doi: 10.1038/ncomms13276.
3
The spike-timing dependence of plasticity.突触传递的时间依赖性可塑性。
Front Neural Circuits. 2025 Aug 14;19:1618506. doi: 10.3389/fncir.2025.1618506. eCollection 2025.
4
Neuromorphic computing for robotic vision: algorithms to hardware advances.用于机器人视觉的神经形态计算:从算法到硬件的进展
Commun Eng. 2025 Aug 13;4(1):152. doi: 10.1038/s44172-025-00492-5.
5
Limits on the computational expressivity of non-equilibrium biophysical processes.非平衡生物物理过程的计算表达能力限制。
Nat Commun. 2025 Aug 5;16(1):7184. doi: 10.1038/s41467-025-61873-0.
6
Quantum equilibrium propagation for efficient training of quantum systems based on Onsager reciprocity.基于昂萨格互易性的量子系统高效训练的量子平衡传播
Nat Commun. 2025 Jul 17;16(1):6595. doi: 10.1038/s41467-025-61665-6.
7
Predictive Coding Model Detects Novelty on Different Levels of Representation Hierarchy.预测编码模型在不同层次的表征层级上检测新颖性。
Neural Comput. 2025 Jul 17;37(8):1373-1408. doi: 10.1162/neco_a_01769.
8
Self-Contrastive Forward-Forward algorithm.自对比前向-前向算法
Nat Commun. 2025 Jul 1;16(1):5978. doi: 10.1038/s41467-025-61037-0.
9
Cellular and subcellular specialization enables biology-constrained deep learning.细胞和亚细胞特化使生物学约束的深度学习成为可能。
bioRxiv. 2025 May 27:2025.05.22.655599. doi: 10.1101/2025.05.22.655599.
10
Temporal Contrastive Learning through implicit non-equilibrium memory.通过隐式非平衡记忆实现时间对比学习。
Nat Commun. 2025 Mar 4;16(1):2163. doi: 10.1038/s41467-025-57043-x.
Neuron. 2012 Aug 23;75(4):556-71. doi: 10.1016/j.neuron.2012.08.001.
4
Spike-timing-dependent plasticity: a comprehensive overview.尖峰时间依赖性可塑性:全面概述。
Front Synaptic Neurosci. 2012 Jul 12;4:2. doi: 10.3389/fnsyn.2012.00002. eCollection 2012.
5
A triplet spike-timing-dependent plasticity model generalizes the Bienenstock-Cooper-Munro rule to higher-order spatiotemporal correlations.三重尖峰时间依赖可塑性模型将 Bienenstock-Cooper-Munro 规则推广到更高阶的时空相关性。
Proc Natl Acad Sci U S A. 2011 Nov 29;108(48):19383-8. doi: 10.1073/pnas.1105933108. Epub 2011 Nov 11.
6
Spontaneous cortical activity reveals hallmarks of an optimal internal model of the environment.自发性皮层活动揭示了环境最佳内部模型的特征。
Science. 2011 Jan 7;331(6013):83-7. doi: 10.1126/science.1195870.
7
Free-energy and the brain.自由能与大脑。
Synthese. 2007 Dec 1;159(3):417-458. doi: 10.1007/s11229-007-9237-y.
8
Nonlinear backpropagation: doing backpropagation without derivatives of the activation function.非线性反向传播:在不使用激活函数导数的情况下进行反向传播。
IEEE Trans Neural Netw. 1997;8(6):1321-7. doi: 10.1109/72.641455.
9
Equivalence of backpropagation and contrastive Hebbian learning in a layered network.分层网络中反向传播与对比赫布学习的等效性。
Neural Comput. 2003 Feb;15(2):441-54. doi: 10.1162/089976603762552988.
10
Training products of experts by minimizing contrastive divergence.通过最小化对比散度来训练专家的产品。
Neural Comput. 2002 Aug;14(8):1771-800. doi: 10.1162/089976602760128018.