• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

无监督学习与聚类连接增强脉冲神经网络中的强化学习

Unsupervised Learning and Clustered Connectivity Enhance Reinforcement Learning in Spiking Neural Networks.

作者信息

Weidel Philipp, Duarte Renato, Morrison Abigail

机构信息

Institute of Neuroscience and Medicine (INM-6) & Institute for Advanced Simulation (IAS-6) & JARA-Institute Brain Structure-Function Relationship (JBI-1 / INM-10), Research Centre Jülich, Jülich, Germany.

Department of Computer Science 3 - Software Engineering, RWTH Aachen University, Aachen, Germany.

出版信息

Front Comput Neurosci. 2021 Mar 4;15:543872. doi: 10.3389/fncom.2021.543872. eCollection 2021.

DOI:10.3389/fncom.2021.543872
PMID:33746728
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7970044/
Abstract

Reinforcement learning is a paradigm that can account for how organisms learn to adapt their behavior in complex environments with sparse rewards. To partition an environment into discrete states, implementations in spiking neuronal networks typically rely on input architectures involving place cells or receptive fields specified by the researcher. This is problematic as a model for how an organism can learn appropriate behavioral sequences in unknown environments, as it fails to account for the unsupervised and self-organized nature of the required representations. Additionally, this approach presupposes knowledge on the part of the researcher on how the environment should be partitioned and represented and scales poorly with the size or complexity of the environment. To address these issues and gain insights into how the brain generates its own task-relevant mappings, we propose a learning architecture that combines unsupervised learning on the input projections with biologically motivated clustered connectivity within the representation layer. This combination allows input features to be mapped to clusters; thus the network self-organizes to produce clearly distinguishable activity patterns that can serve as the basis for reinforcement learning on the output projections. On the basis of the MNIST and Mountain Car tasks, we show that our proposed model performs better than either a comparable unclustered network or a clustered network with static input projections. We conclude that the combination of unsupervised learning and clustered connectivity provides a generic representational substrate suitable for further computation.

摘要

强化学习是一种范式,它可以解释生物体如何在奖励稀疏的复杂环境中学习调整其行为。为了将环境划分为离散状态,脉冲神经网络中的实现通常依赖于涉及位置细胞或由研究人员指定的感受野的输入架构。作为生物体如何在未知环境中学习适当行为序列的模型,这存在问题,因为它没有考虑所需表征的无监督和自组织性质。此外,这种方法预先假定研究人员了解环境应如何划分和表征,并且随着环境的大小或复杂性增加,扩展性较差。为了解决这些问题并深入了解大脑如何生成其自身与任务相关的映射,我们提出了一种学习架构,该架构将输入投影上的无监督学习与表征层内具有生物学动机的聚类连接相结合。这种组合允许将输入特征映射到聚类;因此,网络会自组织以产生清晰可区分的活动模式,这些模式可作为输出投影上强化学习的基础。基于MNIST和山地车任务,我们表明我们提出的模型比可比的非聚类网络或具有静态输入投影的聚类网络表现更好。我们得出结论,无监督学习和聚类连接的组合提供了适合进一步计算的通用表征基础。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bacb/7970044/1e1abc0ec571/fncom-15-543872-g0008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bacb/7970044/72023df91441/fncom-15-543872-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bacb/7970044/a8f447bd80cd/fncom-15-543872-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bacb/7970044/9e2958f78782/fncom-15-543872-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bacb/7970044/ea68cf4a72ae/fncom-15-543872-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bacb/7970044/789e2c90c941/fncom-15-543872-g0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bacb/7970044/e95c5f7a8159/fncom-15-543872-g0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bacb/7970044/c3fe9dd7e6eb/fncom-15-543872-g0007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bacb/7970044/1e1abc0ec571/fncom-15-543872-g0008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bacb/7970044/72023df91441/fncom-15-543872-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bacb/7970044/a8f447bd80cd/fncom-15-543872-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bacb/7970044/9e2958f78782/fncom-15-543872-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bacb/7970044/ea68cf4a72ae/fncom-15-543872-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bacb/7970044/789e2c90c941/fncom-15-543872-g0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bacb/7970044/e95c5f7a8159/fncom-15-543872-g0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bacb/7970044/c3fe9dd7e6eb/fncom-15-543872-g0007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bacb/7970044/1e1abc0ec571/fncom-15-543872-g0008.jpg

相似文献

1
Unsupervised Learning and Clustered Connectivity Enhance Reinforcement Learning in Spiking Neural Networks.无监督学习与聚类连接增强脉冲神经网络中的强化学习
Front Comput Neurosci. 2021 Mar 4;15:543872. doi: 10.3389/fncom.2021.543872. eCollection 2021.
2
SpiLinC: Spiking Liquid-Ensemble Computing for Unsupervised Speech and Image Recognition.SpiLinC:用于无监督语音和图像识别的脉冲液体集成计算
Front Neurosci. 2018 Aug 23;12:524. doi: 10.3389/fnins.2018.00524. eCollection 2018.
3
Biologically plausible deep learning - But how far can we go with shallow networks?生物学上合理的深度学习——但我们可以在浅层网络中走多远?
Neural Netw. 2019 Oct;118:90-101. doi: 10.1016/j.neunet.2019.06.001. Epub 2019 Jun 20.
4
Controlled Forgetting: Targeted Stimulation and Dopaminergic Plasticity Modulation for Unsupervised Lifelong Learning in Spiking Neural Networks.可控遗忘:用于脉冲神经网络中无监督终身学习的靶向刺激与多巴胺能可塑性调制
Front Neurosci. 2020 Jan 28;14:7. doi: 10.3389/fnins.2020.00007. eCollection 2020.
5
Unsupervised learning of digit recognition using spike-timing-dependent plasticity.使用基于脉冲时间依赖可塑性的无监督数字识别学习。
Front Comput Neurosci. 2015 Aug 3;9:99. doi: 10.3389/fncom.2015.00099. eCollection 2015.
6
Locally connected spiking neural networks for unsupervised feature learning.用于无监督特征学习的局部连接脉冲神经网络。
Neural Netw. 2019 Nov;119:332-340. doi: 10.1016/j.neunet.2019.08.016. Epub 2019 Aug 26.
7
Cognitive navigation based on nonuniform Gabor space sampling, unsupervised growing networks, and reinforcement learning.基于非均匀伽柏空间采样、无监督生长网络和强化学习的认知导航。
IEEE Trans Neural Netw. 2004 May;15(3):639-52. doi: 10.1109/TNN.2004.826221.
8
Hierarchical Bayesian Inference and Learning in Spiking Neural Networks.分层贝叶斯推断和尖峰神经网络中的学习。
IEEE Trans Cybern. 2019 Jan;49(1):133-145. doi: 10.1109/TCYB.2017.2768554. Epub 2017 Nov 9.
9
Contextual Integration in Cortical and Convolutional Neural Networks.皮层神经网络和卷积神经网络中的上下文整合
Front Comput Neurosci. 2020 Apr 23;14:31. doi: 10.3389/fncom.2020.00031. eCollection 2020.
10
Reinforcement Learning With Low-Complexity Liquid State Machines.基于低复杂度液态机器的强化学习
Front Neurosci. 2019 Aug 27;13:883. doi: 10.3389/fnins.2019.00883. eCollection 2019.

引用本文的文献

1
An accurate and fast learning approach in the biologically spiking neural network.生物脉冲神经网络中一种准确且快速的学习方法。
Sci Rep. 2025 Feb 24;15(1):6585. doi: 10.1038/s41598-025-90113-0.
2
A spiking neural network for active efficient coding.一种用于主动高效编码的脉冲神经网络。
Front Robot AI. 2025 Jan 15;11:1435197. doi: 10.3389/frobt.2024.1435197. eCollection 2024.
3
Robust self management classification via sparse representation based discriminative model for mild cognitive impairment associated with diabetes mellitus.

本文引用的文献

1
The Interplay of Synaptic Plasticity and Scaling Enables Self-Organized Formation and Allocation of Multiple Memory Representations.突触可塑性和缩放的相互作用使多种记忆表现形式能够自我组织形成和分配。
Front Neural Circuits. 2020 Oct 7;14:541728. doi: 10.3389/fncir.2020.541728. eCollection 2020.
2
Passing the Message: Representation Transfer in Modular Balanced Networks.传递信息:模块化平衡网络中的表示转移
Front Comput Neurosci. 2019 Dec 5;13:79. doi: 10.3389/fncom.2019.00079. eCollection 2019.
3
A deep learning framework for neuroscience.
基于稀疏表示的判别模型对2型糖尿病合并轻度认知障碍 robust self management的分类研究
Sci Rep. 2024 Dec 30;14(1):31779. doi: 10.1038/s41598-024-82665-4.
4
Co-existence of synaptic plasticity and metastable dynamics in a spiking model of cortical circuits.皮质电路尖峰模型中的突触可塑性和亚稳态动力学共存。
PLoS Comput Biol. 2024 Jul 1;20(7):e1012220. doi: 10.1371/journal.pcbi.1012220. eCollection 2024 Jul.
5
Co-existence of synaptic plasticity and metastable dynamics in a spiking model of cortical circuits.皮质回路脉冲模型中突触可塑性与亚稳态动力学的共存
bioRxiv. 2024 Jun 9:2023.12.07.570692. doi: 10.1101/2023.12.07.570692.
6
Long- and short-term history effects in a spiking network model of statistical learning.长短期历史效应在统计学习的尖峰网络模型中的作用。
Sci Rep. 2023 Aug 9;13(1):12939. doi: 10.1038/s41598-023-39108-3.
7
Exploring Parameter and Hyper-Parameter Spaces of Neuroscience Models on High Performance Computers With Learning to Learn.利用学会学习在高性能计算机上探索神经科学模型的参数和超参数空间。
Front Comput Neurosci. 2022 May 27;16:885207. doi: 10.3389/fncom.2022.885207. eCollection 2022.
深度学习在神经科学中的应用框架。
Nat Neurosci. 2019 Nov;22(11):1761-1770. doi: 10.1038/s41593-019-0520-2. Epub 2019 Oct 28.
4
Neural network models and deep learning.神经网络模型与深度学习。
Curr Biol. 2019 Apr 1;29(7):R231-R236. doi: 10.1016/j.cub.2019.02.034.
5
Unsupervised learning by competing hidden units.无监督竞争型隐单元学习。
Proc Natl Acad Sci U S A. 2019 Apr 16;116(16):7723-7731. doi: 10.1073/pnas.1820458116. Epub 2019 Mar 29.
6
Backpropagation through time and the brain.时间反向传播与大脑。
Curr Opin Neurobiol. 2019 Apr;55:82-89. doi: 10.1016/j.conb.2019.01.011. Epub 2019 Mar 7.
7
Theories of Error Back-Propagation in the Brain.大脑中的误差反向传播理论。
Trends Cogn Sci. 2019 Mar;23(3):235-250. doi: 10.1016/j.tics.2018.12.005. Epub 2019 Jan 28.
8
The Interplay Between Cortical State and Perceptual Learning: A Focused Review.皮质状态与知觉学习之间的相互作用:一篇综述
Front Syst Neurosci. 2018 Oct 9;12:47. doi: 10.3389/fnsys.2018.00047. eCollection 2018.
9
Flexible Sensorimotor Computations through Rapid Reconfiguration of Cortical Dynamics.通过快速重新配置皮层动态实现灵活的感觉运动计算。
Neuron. 2018 Jun 6;98(5):1005-1019.e5. doi: 10.1016/j.neuron.2018.05.020.
10
Cortical response states for enhanced sensory discrimination.增强感官辨别力的皮质反应状态。
Elife. 2017 Dec 23;6:e29226. doi: 10.7554/eLife.29226.