• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

预测编码与反向传播之间的关系。

On the relationship between predictive coding and backpropagation.

机构信息

Department of Applied and Computational Mathematics and Statistics, University of Notre Dame, Notre Dame, IN, United States of America.

出版信息

PLoS One. 2022 Mar 31;17(3):e0266102. doi: 10.1371/journal.pone.0266102. eCollection 2022.

DOI:10.1371/journal.pone.0266102
PMID:35358258
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8970408/
Abstract

Artificial neural networks are often interpreted as abstract models of biological neuronal networks, but they are typically trained using the biologically unrealistic backpropagation algorithm and its variants. Predictive coding has been proposed as a potentially more biologically realistic alternative to backpropagation for training neural networks. This manuscript reviews and extends recent work on the mathematical relationship between predictive coding and backpropagation for training feedforward artificial neural networks on supervised learning tasks. Implications of these results for the interpretation of predictive coding and deep neural networks as models of biological learning are discussed along with a repository of functions, Torch2PC, for performing predictive coding with PyTorch neural network models.

摘要

人工神经网络通常被解释为生物神经元网络的抽象模型,但它们通常使用生物上不现实的反向传播算法及其变体进行训练。预测编码被提议作为反向传播的一种潜在更具生物合理性的替代方法,用于训练神经网络。本文综述并扩展了最近关于在监督学习任务上训练前馈人工神经网络的预测编码和反向传播之间的数学关系的工作。讨论了这些结果对预测编码和深度神经网络作为生物学习模型的解释的影响,以及一个用于使用 PyTorch 神经网络模型执行预测编码的函数库 Torch2PC。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/79f8/8970408/8e2f921e571c/pone.0266102.g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/79f8/8970408/d7793132c40c/pone.0266102.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/79f8/8970408/f28e5ba116ed/pone.0266102.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/79f8/8970408/2ed77304dd19/pone.0266102.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/79f8/8970408/e3fc877dee55/pone.0266102.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/79f8/8970408/5e8495718178/pone.0266102.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/79f8/8970408/fbae7b463cd1/pone.0266102.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/79f8/8970408/b4a849ae331d/pone.0266102.g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/79f8/8970408/50882bf98313/pone.0266102.g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/79f8/8970408/8e2f921e571c/pone.0266102.g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/79f8/8970408/d7793132c40c/pone.0266102.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/79f8/8970408/f28e5ba116ed/pone.0266102.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/79f8/8970408/2ed77304dd19/pone.0266102.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/79f8/8970408/e3fc877dee55/pone.0266102.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/79f8/8970408/5e8495718178/pone.0266102.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/79f8/8970408/fbae7b463cd1/pone.0266102.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/79f8/8970408/b4a849ae331d/pone.0266102.g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/79f8/8970408/50882bf98313/pone.0266102.g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/79f8/8970408/8e2f921e571c/pone.0266102.g009.jpg

相似文献

1
On the relationship between predictive coding and backpropagation.预测编码与反向传播之间的关系。
PLoS One. 2022 Mar 31;17(3):e0266102. doi: 10.1371/journal.pone.0266102. eCollection 2022.
2
Biologically plausible deep learning - But how far can we go with shallow networks?生物学上合理的深度学习——但我们可以在浅层网络中走多远?
Neural Netw. 2019 Oct;118:90-101. doi: 10.1016/j.neunet.2019.06.001. Epub 2019 Jun 20.
3
Temporal Coding in Spiking Neural Networks With Alpha Synaptic Function: Learning With Backpropagation.尖峰神经网络中的时间编码与阿尔法突触功能:基于反向传播的学习。
IEEE Trans Neural Netw Learn Syst. 2022 Oct;33(10):5939-5952. doi: 10.1109/TNNLS.2021.3071976. Epub 2022 Oct 5.
4
An Approximation of the Error Backpropagation Algorithm in a Predictive Coding Network with Local Hebbian Synaptic Plasticity.具有局部赫布突触可塑性的预测编码网络中误差反向传播算法的一种近似
Neural Comput. 2017 May;29(5):1229-1262. doi: 10.1162/NECO_a_00949. Epub 2017 Mar 23.
5
Supervised Learning in Neural Networks: Feedback-Network-Free Implementation and Biological Plausibility.神经网络中的监督学习:无反馈网络实现与生物合理性
IEEE Trans Neural Netw Learn Syst. 2022 Dec;33(12):7888-7898. doi: 10.1109/TNNLS.2021.3089134. Epub 2022 Nov 30.
6
Brain-inspired Predictive Coding Improves the Performance of Machine Challenging Tasks.受大脑启发的预测编码提高了机器挑战性任务的性能。
Front Comput Neurosci. 2022 Nov 16;16:1062678. doi: 10.3389/fncom.2022.1062678. eCollection 2022.
7
Deep learning in spiking neural networks.深度学习在尖峰神经网络中的应用。
Neural Netw. 2019 Mar;111:47-63. doi: 10.1016/j.neunet.2018.12.002. Epub 2018 Dec 18.
8
Rectified Linear Postsynaptic Potential Function for Backpropagation in Deep Spiking Neural Networks.深度脉冲神经网络中用于反向传播的整流线性突触后电位函数
IEEE Trans Neural Netw Learn Syst. 2022 May;33(5):1947-1958. doi: 10.1109/TNNLS.2021.3110991. Epub 2022 May 2.
9
A more biologically plausible learning rule than backpropagation applied to a network model of cortical area 7a.一种比反向传播更具生物学合理性的学习规则,应用于皮层7a区的网络模型。
Cereb Cortex. 1991 Jul-Aug;1(4):293-307. doi: 10.1093/cercor/1.4.293.
10
Temporal Backpropagation for Spiking Neural Networks with One Spike per Neuron.神经元每次仅发放一个尖峰的尖峰神经网络的时间反向传播。
Int J Neural Syst. 2020 Jun;30(6):2050027. doi: 10.1142/S0129065720500276. Epub 2020 May 28.

引用本文的文献

1
Effective methods and framework for energy-based local learning of deep neural networks.基于能量的深度神经网络局部学习的有效方法与框架。
Front Artif Intell. 2025 Aug 26;8:1605706. doi: 10.3389/frai.2025.1605706. eCollection 2025.
2
Correction: On the relationship between predictive coding and backpropagation.更正:关于预测编码与反向传播之间的关系。
PLoS One. 2025 Mar 25;20(3):e0320944. doi: 10.1371/journal.pone.0320944. eCollection 2025.
3
Images with harder-to-reconstruct visual representations leave stronger memory traces.

本文引用的文献

1
Predictive Coding Approximates Backprop Along Arbitrary Computation Graphs.预测编码可沿任意计算图逼近反向传播。
Neural Comput. 2022 May 19;34(6):1329-1368. doi: 10.1162/neco_a_01497.
2
Burst-dependent synaptic plasticity can coordinate learning in hierarchical circuits.突发依赖性突触可塑性可以协调分层电路中的学习。
Nat Neurosci. 2021 Jul;24(7):1010-1019. doi: 10.1038/s41593-021-00857-x. Epub 2021 May 13.
3
Can the Brain Do Backpropagation? -Exact Implementation of Backpropagation in Predictive Coding Networks.大脑能进行反向传播吗?——预测编码网络中反向传播的精确实现。
视觉再现较难重构的图像会留下更强的记忆痕迹。
Nat Hum Behav. 2024 Jul;8(7):1309-1320. doi: 10.1038/s41562-024-01870-3. Epub 2024 May 13.
4
An intelligent diabetes classification and perception framework based on ensemble and deep learning method.基于集成与深度学习方法的智能糖尿病分类与感知框架
PeerJ Comput Sci. 2024 Mar 29;10:e1914. doi: 10.7717/peerj-cs.1914. eCollection 2024.
5
Surprisal From Language Models Can Predict ERPs in Processing Predicate-Argument Structures Only if Enriched by an Agent Preference Principle.只有通过主体偏好原则进行强化,语言模型的意外值才能预测谓词-论元结构处理过程中的事件相关电位。
Neurobiol Lang (Camb). 2024 Apr 1;5(1):167-200. doi: 10.1162/nol_a_00121. eCollection 2024.
6
Learning cortical hierarchies with temporal Hebbian updates.通过时间赫布更新学习皮层层次结构。
Front Comput Neurosci. 2023 May 24;17:1136010. doi: 10.3389/fncom.2023.1136010. eCollection 2023.
7
Recurrent predictive coding models for associative memory employing covariance learning.采用协方差学习的联想记忆反复预测编码模型。
PLoS Comput Biol. 2023 Apr 14;19(4):e1010719. doi: 10.1371/journal.pcbi.1010719. eCollection 2023 Apr.
8
Brain-inspired Predictive Coding Improves the Performance of Machine Challenging Tasks.受大脑启发的预测编码提高了机器挑战性任务的性能。
Front Comput Neurosci. 2022 Nov 16;16:1062678. doi: 10.3389/fncom.2022.1062678. eCollection 2022.
9
Prediction-error neurons in circuits with multiple neuron types: Formation, refinement, and functional implications.具有多种神经元类型的回路中的预测误差神经元:形成、细化和功能意义。
Proc Natl Acad Sci U S A. 2022 Mar 29;119(13):e2115699119. doi: 10.1073/pnas.2115699119. Epub 2022 Mar 23.
Adv Neural Inf Process Syst. 2020;33:22566-22579.
4
Integrative Benchmarking to Advance Neurally Mechanistic Models of Human Intelligence.综合基准测试以推进人类智能的神经机制模型。
Neuron. 2020 Nov 11;108(3):413-423. doi: 10.1016/j.neuron.2020.07.040. Epub 2020 Sep 11.
5
Learning prediction error neurons in a canonical interneuron circuit.学习规范中间神经元回路中的预测误差神经元。
Elife. 2020 Aug 21;9:e57541. doi: 10.7554/eLife.57541.
6
Backpropagation and the brain.反向传播与大脑。
Nat Rev Neurosci. 2020 Jun;21(6):335-346. doi: 10.1038/s41583-020-0277-3. Epub 2020 Apr 17.
7
Theories of Error Back-Propagation in the Brain.大脑中的误差反向传播理论。
Trends Cogn Sci. 2019 Mar;23(3):235-250. doi: 10.1016/j.tics.2018.12.005. Epub 2019 Jan 28.
8
Predictive Processing: A Canonical Cortical Computation.预测加工:一种经典的皮层计算。
Neuron. 2018 Oct 24;100(2):424-435. doi: 10.1016/j.neuron.2018.10.003.
9
Visuomotor Coupling Shapes the Functional Development of Mouse Visual Cortex.视动耦合塑造了小鼠视觉皮层的功能发育。
Cell. 2017 Jun 15;169(7):1291-1302.e14. doi: 10.1016/j.cell.2017.05.023. Epub 2017 Jun 9.
10
Equilibrium Propagation: Bridging the Gap between Energy-Based Models and Backpropagation.平衡传播:弥合基于能量模型与反向传播之间的差距
Front Comput Neurosci. 2017 May 4;11:24. doi: 10.3389/fncom.2017.00024. eCollection 2017.