• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

利用聚合测量学习控制神经元

Learning to Control Neurons using Aggregated Measurements.

作者信息

Yu Yao-Chi, Narayanan Vignesh, Ching ShiNung, Li Jr-Shin

机构信息

Department of Electrical and Systems Engineering, Washington University in St. Louis, St. Louis, MO, 63130, USA.

Division of Biology and Biomedical Sciences, Washington University in St. Louis, St. Louis, MO, 63130, USA.

出版信息

Proc Am Control Conf. 2020 Jul;2020:4028-4033. doi: 10.23919/acc45564.2020.9147426. Epub 2020 Jul 27.

DOI:10.23919/acc45564.2020.9147426
PMID:38009125
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10676147/
Abstract

Controlling a population of neurons with one or a few control signals is challenging due to the severely underactuated nature of the control system and the inherent nonlinear dynamics of the neurons that are typically unknown. Control strategies that incorporate deep neural networks and machine learning techniques directly use data to learn a sequence of control actions for targeted manipulation of a population of neurons. However, these learning strategies inherently assume that perfect feedback data from each neuron at every sampling instant are available, and do not scale gracefully as the number of neurons in the population increases. As a result, the learning models need to be retrained whenever such a change occurs. In this work, we propose a learning strategy to design a control sequence by using population-level aggregated measurements and incorporate reinforcement learning techniques to find a (bounded, piecewise constant) control policy that fulfills the given control task. We demonstrate the feasibility of the proposed approach using numerical experiments on a finite population of nonlinear dynamical systems and canonical phase models that are widely used in neuroscience.

摘要

由于控制系统严重欠驱动的特性以及通常未知的神经元固有非线性动力学,用一个或几个控制信号来控制一群神经元具有挑战性。结合深度神经网络和机器学习技术的控制策略直接利用数据来学习一系列控制动作,以对一群神经元进行有针对性的操纵。然而,这些学习策略本质上假设在每个采样时刻都能获得来自每个神经元的完美反馈数据,并且随着群体中神经元数量的增加,其扩展性不佳。因此,每当发生这种变化时,学习模型都需要重新训练。在这项工作中,我们提出一种学习策略,通过使用群体水平的聚合测量来设计控制序列,并结合强化学习技术来找到一个满足给定控制任务的(有界、分段常数)控制策略。我们通过对有限数量的非线性动力系统和神经科学中广泛使用的典型相位模型进行数值实验,证明了所提出方法的可行性。

相似文献

1
Learning to Control Neurons using Aggregated Measurements.利用聚合测量学习控制神经元
Proc Am Control Conf. 2020 Jul;2020:4028-4033. doi: 10.23919/acc45564.2020.9147426. Epub 2020 Jul 27.
2
Moment-Based Reinforcement Learning for Ensemble Control.用于集成控制的基于矩的强化学习
IEEE Trans Neural Netw Learn Syst. 2024 Sep;35(9):12653-12664. doi: 10.1109/TNNLS.2023.3264151. Epub 2024 Sep 4.
3
Reinforcement Learning for Central Pattern Generation in Dynamical Recurrent Neural Networks.用于动态递归神经网络中中枢模式生成的强化学习
Front Comput Neurosci. 2022 Apr 8;16:818985. doi: 10.3389/fncom.2022.818985. eCollection 2022.
4
Fast deep neural correspondence for tracking and identifying neurons in using semi-synthetic training.利用半合成训练进行快速深层神经对应,以跟踪和识别 中的神经元。
Elife. 2021 Jul 14;10:e66410. doi: 10.7554/eLife.66410.
5
PaCAR: COVID-19 Pandemic Control Decision Making via Large-Scale Agent-Based Modeling and Deep Reinforcement Learning.PaCAR:通过大规模基于代理的建模和深度强化学习进行 COVID-19 大流行控制决策。
Med Decis Making. 2022 Nov;42(8):1064-1077. doi: 10.1177/0272989X221107902. Epub 2022 Jul 1.
6
In Vivo Observations of Rapid Scattered Light Changes Associated with Neurophysiological Activity与神经生理活动相关的快速散射光变化的体内观察
7
Event-Driven Off-Policy Reinforcement Learning for Control of Interconnected Systems.事件驱动的非策略强化学习在互联系统控制中的应用。
IEEE Trans Cybern. 2022 Mar;52(3):1936-1946. doi: 10.1109/TCYB.2020.2991166. Epub 2022 Mar 11.
8
Evolving interpretable plasticity for spiking networks.用于脉冲神经网络的不断发展的可解释可塑性。
Elife. 2021 Oct 28;10:e66273. doi: 10.7554/eLife.66273.
9
Discrete-time online learning control for a class of unknown nonaffine nonlinear systems using reinforcement learning.基于强化学习的一类未知非仿射非线性系统的离散时间在线学习控制。
Neural Netw. 2014 Jul;55:30-41. doi: 10.1016/j.neunet.2014.03.008. Epub 2014 Mar 28.
10
Dynamic event-triggered controller design for nonlinear systems: Reinforcement learning strategy.动态事件触发控制器设计的非线性系统:强化学习策略。
Neural Netw. 2023 Jun;163:341-353. doi: 10.1016/j.neunet.2023.04.008. Epub 2023 Apr 19.

引用本文的文献

1
Data-Driven Control of Neuronal Networks with Population-Level Measurement.基于群体水平测量的神经网络数据驱动控制
Res Sq. 2023 Mar 17:rs.3.rs-2600572. doi: 10.21203/rs.3.rs-2600572/v1.

本文引用的文献

1
Learning-based Approaches for Controlling Neural Spiking.基于学习的神经放电控制方法。
Proc Am Control Conf. 2018 Jun;2018. doi: 10.23919/acc.2018.8431158. Epub 2018 Aug 16.
2
Model Learning and Knowledge Sharing for Cooperative Multiagent Systems in Stochastic Environment.随机环境下合作多智能体系统的模型学习与知识共享
IEEE Trans Cybern. 2021 Dec;51(12):5717-5727. doi: 10.1109/TCYB.2019.2958912. Epub 2021 Dec 22.
3
Novel electrode technologies for neural recordings.新型神经记录电极技术。
Nat Rev Neurosci. 2019 Jun;20(6):330-345. doi: 10.1038/s41583-019-0140-6.
4
Control of neural systems at multiple scales using model-free, deep reinforcement learning.使用无模型、深度强化学习控制多尺度神经系统。
Sci Rep. 2018 Jul 16;8(1):10721. doi: 10.1038/s41598-018-29134-x.
5
Fundamental Limits of Forced Asynchronous Spiking with Integrate and Fire Dynamics.基于积分发放动力学的强迫异步尖峰的基本限制
J Math Neurosci. 2017 Oct 11;7(1):11. doi: 10.1186/s13408-017-0053-5.
6
Designing optimal stimuli to control neuronal spike timing.设计最优刺激来控制神经元尖峰时间。
J Neurophysiol. 2011 Aug;106(2):1038-53. doi: 10.1152/jn.00427.2010. Epub 2011 Apr 20.
7
Single-neuron dynamics in human focal epilepsy.人类局灶性癫痫中单神经元动力学。
Nat Neurosci. 2011 May;14(5):635-41. doi: 10.1038/nn.2782. Epub 2011 Mar 27.
8
Noise in the nervous system.神经系统中的噪音。
Nat Rev Neurosci. 2008 Apr;9(4):292-303. doi: 10.1038/nrn2258.
9
Deep brain stimulation for dystonia.用于治疗肌张力障碍的深部脑刺激术。
J Clin Neurophysiol. 2004 Jan-Feb;21(1):18-30. doi: 10.1097/00004691-200401000-00004.
10
On the phase reduction and response dynamics of neural oscillator populations.关于神经振荡器群体的相位简化与响应动力学
Neural Comput. 2004 Apr;16(4):673-715. doi: 10.1162/089976604322860668.