• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

通过脑机接口区分学习规则

Distinguishing Learning Rules with Brain Machine Interfaces.

作者信息

Portes Jacob P, Schmid Christian, Murray James M

机构信息

Center for Theoretical Neuroscience, Columbia University.

Institute of Neuroscience, University of Oregon.

出版信息

Adv Neural Inf Process Syst. 2022 Dec;35:25937-25950.

PMID:37101843
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10129057/
Abstract

Despite extensive theoretical work on biologically plausible learning rules, clear evidence about whether and how such rules are implemented in the brain has been difficult to obtain. We consider biologically plausible supervised- and reinforcement-learning rules and ask whether changes in network activity during learning can be used to determine which learning rule is being used. Supervised learning requires a credit-assignment model estimating the mapping from neural activity to behavior, and, in a biological organism, this model will inevitably be an imperfect approximation of the ideal mapping, leading to a bias in the direction of the weight updates relative to the true gradient. Reinforcement learning, on the other hand, requires no credit-assignment model and tends to make weight updates following the true gradient direction. We derive a metric to distinguish between learning rules by observing changes in the network activity during learning, given that the mapping from brain to behavior is known by the experimenter. Because brain-machine interface (BMI) experiments allow for precise knowledge of this mapping, we model a cursor-control BMI task using recurrent neural networks, showing that learning rules can be distinguished in simulated experiments using only observations that a neuroscience experimenter would plausibly have access to.

摘要

尽管在具有生物学合理性的学习规则方面已经开展了大量理论研究工作,但关于此类规则是否以及如何在大脑中得以实现的明确证据却一直难以获得。我们考虑具有生物学合理性的监督学习和强化学习规则,并探究学习过程中网络活动的变化是否可用于确定正在使用的是哪种学习规则。监督学习需要一个信用分配模型来估计从神经活动到行为的映射,而在生物有机体中,该模型不可避免地是理想映射的不完美近似,从而导致相对于真实梯度的权重更新方向出现偏差。另一方面,强化学习不需要信用分配模型,并且倾向于沿着真实梯度方向进行权重更新。鉴于实验者已知从大脑到行为的映射,我们通过观察学习过程中网络活动的变化来推导一种区分学习规则的度量。由于脑机接口(BMI)实验能够精确了解这种映射,我们使用循环神经网络对光标控制BMI任务进行建模,结果表明,在模拟实验中,仅使用神经科学实验者可能获取的观察结果就能区分学习规则。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/58cc/10129057/a4d84b078539/nihms-1843275-f0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/58cc/10129057/1c5a6f610cfd/nihms-1843275-f0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/58cc/10129057/a4cc328df16e/nihms-1843275-f0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/58cc/10129057/336782dcd9ed/nihms-1843275-f0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/58cc/10129057/fa7b87edcd02/nihms-1843275-f0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/58cc/10129057/a4d84b078539/nihms-1843275-f0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/58cc/10129057/1c5a6f610cfd/nihms-1843275-f0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/58cc/10129057/a4cc328df16e/nihms-1843275-f0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/58cc/10129057/336782dcd9ed/nihms-1843275-f0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/58cc/10129057/fa7b87edcd02/nihms-1843275-f0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/58cc/10129057/a4d84b078539/nihms-1843275-f0005.jpg

相似文献

1
Distinguishing Learning Rules with Brain Machine Interfaces.通过脑机接口区分学习规则
Adv Neural Inf Process Syst. 2022 Dec;35:25937-25950.
2
Biologically plausible local synaptic learning rules robustly implement deep supervised learning.具有生物学合理性的局部突触学习规则有力地实现了深度监督学习。
Front Neurosci. 2023 Oct 11;17:1160899. doi: 10.3389/fnins.2023.1160899. eCollection 2023.
3
Attention-gated reinforcement learning of internal representations for classification.用于分类的内部表征的注意力门控强化学习。
Neural Comput. 2005 Oct;17(10):2176-214. doi: 10.1162/0899766054615699.
4
A biologically plausible supervised learning method for spiking neural networks using the symmetric STDP rule.基于对称 STDP 规则的尖峰神经网络的生物合理有监督学习方法。
Neural Netw. 2020 Jan;121:387-395. doi: 10.1016/j.neunet.2019.09.007. Epub 2019 Sep 27.
5
Target spike patterns enable efficient and biologically plausible learning for complex temporal tasks.目标尖峰模式可实现高效且在生物学上合理的复杂时间任务学习。
PLoS One. 2021 Feb 16;16(2):e0247014. doi: 10.1371/journal.pone.0247014. eCollection 2021.
6
Intermediate Sensory Feedback Assisted Multi-Step Neural Decoding for Reinforcement Learning Based Brain-Machine Interfaces.基于强化学习的脑机接口的中间感觉反馈辅助多步神经解码。
IEEE Trans Neural Syst Rehabil Eng. 2022;30:2834-2844. doi: 10.1109/TNSRE.2022.3210700. Epub 2022 Oct 20.
7
Neuromodulated Synaptic Plasticity on the SpiNNaker Neuromorphic System.基于SpiNNaker神经形态系统的神经调节突触可塑性
Front Neurosci. 2018 Feb 27;12:105. doi: 10.3389/fnins.2018.00105. eCollection 2018.
8
Statistical mechanics of structural and temporal credit assignment effects on learning in neural networks.神经网络中结构和时间信用分配对学习影响的统计力学
Phys Rev E Stat Nonlin Soft Matter Phys. 2011 May;83(5 Pt 1):051125. doi: 10.1103/PhysRevE.83.051125. Epub 2011 May 20.
9
Reinforcement Learning for Central Pattern Generation in Dynamical Recurrent Neural Networks.用于动态递归神经网络中中枢模式生成的强化学习
Front Comput Neurosci. 2022 Apr 8;16:818985. doi: 10.3389/fncom.2022.818985. eCollection 2022.
10
Can the Brain Do Backpropagation? -Exact Implementation of Backpropagation in Predictive Coding Networks.大脑能进行反向传播吗?——预测编码网络中反向传播的精确实现。
Adv Neural Inf Process Syst. 2020;33:22566-22579.

引用本文的文献

1
Model-based inference of synaptic plasticity rules.基于模型的突触可塑性规则推理。
Adv Neural Inf Process Syst. 2024;37:48519-48540.
2
Dynamics of Supervised and Reinforcement Learning in the Non-Linear Perceptron.非线性感知器中监督学习与强化学习的动态变化
ArXiv. 2025 Feb 24:arXiv:2409.03749v3.

本文引用的文献

1
Small, correlated changes in synaptic connectivity may facilitate rapid motor learning.突触连接的微小、相关变化可能有助于快速运动学习。
Nat Commun. 2022 Sep 2;13(1):5163. doi: 10.1038/s41467-022-32646-w.
2
How learning unfolds in the brain: toward an optimization view.学习在大脑中是如何展开的:走向优化的观点。
Neuron. 2021 Dec 1;109(23):3720-3735. doi: 10.1016/j.neuron.2021.09.005. Epub 2021 Oct 13.
3
Neural manifold under plasticity in a goal driven learning behaviour.在目标驱动的学习行为中,神经流形下的可塑性。
PLoS Comput Biol. 2021 Feb 5;17(2):e1008621. doi: 10.1371/journal.pcbi.1008621. eCollection 2021 Feb.
4
Backpropagation and the brain.反向传播与大脑。
Nat Rev Neurosci. 2020 Jun;21(6):335-346. doi: 10.1038/s41583-020-0277-3. Epub 2020 Apr 17.
5
Neural reinforcement: re-entering and refining neural dynamics leading to desirable outcomes.神经强化:重新进入和细化导致期望结果的神经动力学。
Curr Opin Neurobiol. 2020 Feb;60:145-154. doi: 10.1016/j.conb.2019.11.023. Epub 2019 Dec 24.
6
A deep learning framework for neuroscience.深度学习在神经科学中的应用框架。
Nat Neurosci. 2019 Nov;22(11):1761-1770. doi: 10.1038/s41593-019-0520-2. Epub 2019 Oct 28.
7
New neural activity patterns emerge with long-term learning.新的神经活动模式随着长期学习而出现。
Proc Natl Acad Sci U S A. 2019 Jul 23;116(30):15210-15215. doi: 10.1073/pnas.1820296116. Epub 2019 Jun 10.
8
Local online learning in recurrent networks with random feedback.具有随机反馈的递归网络中的局部在线学习。
Elife. 2019 May 24;8:e43299. doi: 10.7554/eLife.43299.
9
Deep Learning With Asymmetric Connections and Hebbian Updates.具有非对称连接和赫布更新的深度学习
Front Comput Neurosci. 2019 Apr 4;13:18. doi: 10.3389/fncom.2019.00018. eCollection 2019.
10
Motor Learning.运动学习。
Compr Physiol. 2019 Mar 14;9(2):613-663. doi: 10.1002/cphy.c170043.