• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于神经解码的核时间差异

Kernel temporal differences for neural decoding.

作者信息

Bae Jihye, Sanchez Giraldo Luis G, Pohlmeyer Eric A, Francis Joseph T, Sanchez Justin C, Príncipe José C

机构信息

Department of Electrical and Computer Engineering, University of Florida, Gainesville, FL 32611, USA.

Department of Biomedical Engineering, University of Miami, Coral Gables, FL 33146, USA.

出版信息

Comput Intell Neurosci. 2015;2015:481375. doi: 10.1155/2015/481375. Epub 2015 Mar 17.

DOI:10.1155/2015/481375
PMID:25866504
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC4381863/
Abstract

We study the feasibility and capability of the kernel temporal difference (KTD)(λ) algorithm for neural decoding. KTD(λ) is an online, kernel-based learning algorithm, which has been introduced to estimate value functions in reinforcement learning. This algorithm combines kernel-based representations with the temporal difference approach to learning. One of our key observations is that by using strictly positive definite kernels, algorithm's convergence can be guaranteed for policy evaluation. The algorithm's nonlinear functional approximation capabilities are shown in both simulations of policy evaluation and neural decoding problems (policy improvement). KTD can handle high-dimensional neural states containing spatial-temporal information at a reasonable computational complexity allowing real-time applications. When the algorithm seeks a proper mapping between a monkey's neural states and desired positions of a computer cursor or a robot arm, in both open-loop and closed-loop experiments, it can effectively learn the neural state to action mapping. Finally, a visualization of the coadaptation process between the decoder and the subject shows the algorithm's capabilities in reinforcement learning brain machine interfaces.

摘要

我们研究了内核时间差分(KTD)(λ)算法用于神经解码的可行性和能力。KTD(λ)是一种基于内核的在线学习算法,已被引入到强化学习中用于估计价值函数。该算法将基于内核的表示与时间差分学习方法相结合。我们的一个关键发现是,通过使用严格正定内核,可以保证算法在策略评估中的收敛性。该算法的非线性函数逼近能力在策略评估和神经解码问题(策略改进)的模拟中均有体现。KTD能够以合理的计算复杂度处理包含时空信息的高维神经状态,从而实现实时应用。当该算法在开环和闭环实验中寻求猴子神经状态与计算机光标或机器人手臂期望位置之间的适当映射时,它能够有效地学习神经状态到动作的映射。最后,解码器与受试者之间共适应过程的可视化展示了该算法在强化学习脑机接口中的能力。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/73dc7ed88317/CIN2015-481375.015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/c15cf99596be/CIN2015-481375.001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/c48582b001c6/CIN2015-481375.002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/9c3aed855b35/CIN2015-481375.003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/3f9ecfc64a6e/CIN2015-481375.004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/f4578b3ffcaf/CIN2015-481375.005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/63dd216bc8e0/CIN2015-481375.006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/967797e960b5/CIN2015-481375.007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/b5191cc060f8/CIN2015-481375.008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/832bfda7469a/CIN2015-481375.009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/02724f229ce0/CIN2015-481375.010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/920050cb253d/CIN2015-481375.011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/b174995129f9/CIN2015-481375.012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/93cbdd3b8e74/CIN2015-481375.013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/6c6f3396da96/CIN2015-481375.014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/73dc7ed88317/CIN2015-481375.015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/c15cf99596be/CIN2015-481375.001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/c48582b001c6/CIN2015-481375.002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/9c3aed855b35/CIN2015-481375.003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/3f9ecfc64a6e/CIN2015-481375.004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/f4578b3ffcaf/CIN2015-481375.005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/63dd216bc8e0/CIN2015-481375.006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/967797e960b5/CIN2015-481375.007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/b5191cc060f8/CIN2015-481375.008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/832bfda7469a/CIN2015-481375.009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/02724f229ce0/CIN2015-481375.010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/920050cb253d/CIN2015-481375.011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/b174995129f9/CIN2015-481375.012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/93cbdd3b8e74/CIN2015-481375.013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/6c6f3396da96/CIN2015-481375.014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d7ff/4381863/73dc7ed88317/CIN2015-481375.015.jpg

相似文献

1
Kernel temporal differences for neural decoding.用于神经解码的核时间差异
Comput Intell Neurosci. 2015;2015:481375. doi: 10.1155/2015/481375. Epub 2015 Mar 17.
2
Kernel Temporal Differences for EEG-based Reinforcement Learning Brain Machine Interfaces.基于脑电的强化学习脑机接口的核时频差。
Annu Int Conf IEEE Eng Med Biol Soc. 2022 Jul;2022:3327-3333. doi: 10.1109/EMBC48229.2022.9871862.
3
A new method of concurrently visualizing states, values, and actions in reinforcement based brain machine interfaces.一种在基于强化学习的脑机接口中同时可视化状态、值和动作的新方法。
Annu Int Conf IEEE Eng Med Biol Soc. 2013;2013:5402-5. doi: 10.1109/EMBC.2013.6610770.
4
Clustering Neural Patterns in Kernel Reinforcement Learning Assists Fast Brain Control in Brain-Machine Interfaces.核强化学习中的神经模式聚类有助于脑机接口中的快速脑控。
IEEE Trans Neural Syst Rehabil Eng. 2019 Sep;27(9):1684-1694. doi: 10.1109/TNSRE.2019.2934176. Epub 2019 Aug 9.
5
Kernel-based least squares policy iteration for reinforcement learning.用于强化学习的基于核的最小二乘策略迭代
IEEE Trans Neural Netw. 2007 Jul;18(4):973-92. doi: 10.1109/TNN.2007.899161.
6
Reinforcement learning via kernel temporal difference.基于核时间差分的强化学习
Annu Int Conf IEEE Eng Med Biol Soc. 2011;2011:5662-5. doi: 10.1109/IEMBS.2011.6091370.
7
A Kernel Reinforcement Learning Decoding Framework Integrating Neural and Feedback Signals for Brain Control.一种整合神经和反馈信号的核强化学习解码框架,用于脑控。
Annu Int Conf IEEE Eng Med Biol Soc. 2023 Jul;2023:1-4. doi: 10.1109/EMBC40787.2023.10340203.
8
Entropy-based kernel mixture modeling for topographic map formation.用于地形图形成的基于熵的核混合建模
IEEE Trans Neural Netw. 2004 Jul;15(4):850-8. doi: 10.1109/TNN.2004.828763.
9
Local-learning-based neuron selection for grasping gesture prediction in motor brain machine interfaces.基于局部学习的神经元选择在运动脑机接口中的抓取手势预测。
J Neural Eng. 2013 Apr;10(2):026008. doi: 10.1088/1741-2560/10/2/026008. Epub 2013 Feb 21.
10
Intermediate Sensory Feedback Assisted Multi-Step Neural Decoding for Reinforcement Learning Based Brain-Machine Interfaces.基于强化学习的脑机接口的中间感觉反馈辅助多步神经解码。
IEEE Trans Neural Syst Rehabil Eng. 2022;30:2834-2844. doi: 10.1109/TNSRE.2022.3210700. Epub 2022 Oct 20.

引用本文的文献

1
Neural Decoders Using Reinforcement Learning in Brain Machine Interfaces: A Technical Review.脑机接口中使用强化学习的神经解码器:技术综述
Front Syst Neurosci. 2022 Aug 26;16:836778. doi: 10.3389/fnsys.2022.836778. eCollection 2022.
2
Kernel Recursive Least-Squares Temporal Difference Algorithms with Sparsification and Regularization.具有稀疏化和正则化的核递归最小二乘时间差分算法
Comput Intell Neurosci. 2016;2016:2305854. doi: 10.1155/2016/2305854. Epub 2016 Jun 29.

本文引用的文献

1
Quantized kernel least mean square algorithm.量化核最小均方算法。
IEEE Trans Neural Netw Learn Syst. 2012 Jan;23(1):22-32. doi: 10.1109/TNNLS.2011.2178446.
2
Online selective kernel-based temporal difference learning.在线选择性核时变差分学习。
IEEE Trans Neural Netw Learn Syst. 2013 Dec;24(12):1944-56. doi: 10.1109/TNNLS.2013.2270561.
3
Using reinforcement learning to provide stable brain-machine interface control despite neural input reorganization.利用强化学习在神经输入重组的情况下提供稳定的脑机接口控制。
PLoS One. 2014 Jan 30;9(1):e87253. doi: 10.1371/journal.pone.0087253. eCollection 2014.
4
Probabilistic co-adaptive brain-computer interfacing.概率协同自适应脑机接口。
J Neural Eng. 2013 Dec;10(6):066008. doi: 10.1088/1741-2560/10/6/066008. Epub 2013 Oct 18.
5
A new method of concurrently visualizing states, values, and actions in reinforcement based brain machine interfaces.一种在基于强化学习的脑机接口中同时可视化状态、值和动作的新方法。
Annu Int Conf IEEE Eng Med Biol Soc. 2013;2013:5402-5. doi: 10.1109/EMBC.2013.6610770.
6
Reinforcement learning via kernel temporal difference.基于核时间差分的强化学习
Annu Int Conf IEEE Eng Med Biol Soc. 2011;2011:5662-5. doi: 10.1109/IEMBS.2011.6091370.
7
An information theoretic approach of designing sparse kernel adaptive filters.一种设计稀疏核自适应滤波器的信息论方法。
IEEE Trans Neural Netw. 2009 Dec;20(12):1950-61. doi: 10.1109/TNN.2009.2033676. Epub 2009 Nov 17.
8
Coadaptive brain-machine interface via reinforcement learning.通过强化学习实现的协同自适应脑机接口
IEEE Trans Biomed Eng. 2009 Jan;56(1):54-64. doi: 10.1109/TBME.2008.926699.