• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于隐藏脑状态的脑机接口内部评估:使用核逆强化学习方法

Hidden Brain State-based Internal Evaluation Using Kernel Inverse Reinforcement Learning in Brain-machine Interfaces.

作者信息

Tan Jieyuan, Zhang Xiang, Wu Shenghui, Song Zhiwei, Wang Yiwen

出版信息

IEEE Trans Neural Syst Rehabil Eng. 2024 Nov 21;PP. doi: 10.1109/TNSRE.2024.3503713.

DOI:10.1109/TNSRE.2024.3503713
PMID:40030403
Abstract

Reinforcement learning (RL)-based brain machine interfaces (BMIs) assist paralyzed people in controlling neural prostheses without the need for real limb movement as supervised signals. The design of reward signal significantly impacts the learning efficiency of the RL-based decoders. Existing reward designs in the RL-based BMI framework rely on external rewards or manually labeled internal rewards, unable to accurately extract subjects' internal evaluation. In this paper, we propose a hidden brain state-based kernel inverse reinforcement learning (HBS-KIRL) method to accurately infer the subject-specific internal evaluation from neural activity during the BMI task. The state-space model is applied to project the neural state into low-dimensional hidden brain state space, which greatly reduces the exploration dimension. Then the kernel method is applied to speed up the convergence of policy, reward, and Q-value networks in reproducing kernel Hilbert space (RKHS). We tested our proposed algorithm on the data collected from the medial prefrontal cortex (mPFC) of rats when they were performing a two-lever-discrimination task. We assessed the state-value estimation performance of our proposed method and compared it with naïve IRL and PCA-based IRL. To validate that the extracted internal evaluation could contribute to the decoder training, we compared the decoding performance of decoders trained by different reward models, including manually designed reward, naïve IRL, PCA-IRL, and our proposed HBS-KIRL. The results show that the HBS-KIRL method can give a stable and accurate estimation of state-value distribution with respect to behavior. Compared with other methods, the decoder guided by HBS-KIRL achieves consistent and better decoding performance over days. This study reveals the potential of applying the IRL method to better extract subject-specific evaluation and improve the BMI decoding performance.

摘要

基于强化学习(RL)的脑机接口(BMI)可帮助瘫痪患者控制神经假体,而无需真实肢体运动作为监督信号。奖励信号的设计对基于RL的解码器的学习效率有显著影响。基于RL的BMI框架中现有的奖励设计依赖于外部奖励或人工标注的内部奖励,无法准确提取受试者的内部评价。在本文中,我们提出了一种基于隐藏脑状态的核逆强化学习(HBS-KIRL)方法,以从BMI任务期间的神经活动中准确推断特定于受试者的内部评价。应用状态空间模型将神经状态投影到低维隐藏脑状态空间,这大大降低了探索维度。然后应用核方法加速策略、奖励和Q值网络在再生核希尔伯特空间(RKHS)中的收敛。我们在大鼠执行双杠杆辨别任务时从其内侧前额叶皮层(mPFC)收集的数据上测试了我们提出的算法。我们评估了我们提出的方法的状态值估计性能,并将其与朴素逆强化学习(IRL)和基于主成分分析(PCA)的IRL进行了比较。为了验证提取的内部评价有助于解码器训练,我们比较了由不同奖励模型训练的解码器的解码性能,包括人工设计的奖励、朴素IRL、PCA-IRL和我们提出的HBS-KIRL。结果表明,HBS-KIRL方法能够给出关于行为的稳定且准确的状态值分布估计。与其他方法相比,由HBS-KIRL引导的解码器在数天内实现了一致且更好的解码性能。这项研究揭示了应用IRL方法更好地提取特定于受试者的评价并提高BMI解码性能的潜力。

相似文献

1
Hidden Brain State-based Internal Evaluation Using Kernel Inverse Reinforcement Learning in Brain-machine Interfaces.基于隐藏脑状态的脑机接口内部评估:使用核逆强化学习方法
IEEE Trans Neural Syst Rehabil Eng. 2024 Nov 21;PP. doi: 10.1109/TNSRE.2024.3503713.
2
Estimating Reward Function from Medial Prefrontal Cortex Cortical Activity using Inverse Reinforcement Learning.基于逆强化学习从内侧前额叶皮质皮层活动估算奖励函数。
Annu Int Conf IEEE Eng Med Biol Soc. 2022 Jul;2022:3346-3349. doi: 10.1109/EMBC48229.2022.9871194.
3
State-space Model Based Inverse Reinforcement Learning for Reward Function Estimation in Brain-machine Interfaces.基于状态空间模型的脑机接口奖励函数估计逆强化学习
Annu Int Conf IEEE Eng Med Biol Soc. 2023 Jul;2023:1-4. doi: 10.1109/EMBC40787.2023.10340953.
4
Intermediate Sensory Feedback Assisted Multi-Step Neural Decoding for Reinforcement Learning Based Brain-Machine Interfaces.基于强化学习的脑机接口的中间感觉反馈辅助多步神经解码。
IEEE Trans Neural Syst Rehabil Eng. 2022;30:2834-2844. doi: 10.1109/TNSRE.2022.3210700. Epub 2022 Oct 20.
5
Task Learning Over Multi-Day Recording via Internally Rewarded Reinforcement Learning Based Brain Machine Interfaces.基于内部奖励强化学习的多日记录任务学习脑机接口。
IEEE Trans Neural Syst Rehabil Eng. 2020 Dec;28(12):3089-3099. doi: 10.1109/TNSRE.2020.3039970. Epub 2021 Jan 28.
6
Dynamic Inverse Reinforcement Learning for Feedback-driven Reward Estimation in Brain Machine Interface Tasks.用于脑机接口任务中反馈驱动奖励估计的动态逆强化学习
Annu Int Conf IEEE Eng Med Biol Soc. 2024 Jul;2024:1-4. doi: 10.1109/EMBC53108.2024.10782800.
7
Reinforcement Learning based Decoding Using Internal Reward for Time Delayed Task in Brain Machine Interfaces.基于强化学习的解码:利用内部奖励实现脑机接口中的时延任务
Annu Int Conf IEEE Eng Med Biol Soc. 2020 Jul;2020:3351-3354. doi: 10.1109/EMBC44109.2020.9175964.
8
Cluster Kernel Reinforcement Learning-based Kalman Filter for Three-Lever Discrimination Task in Brain-Machine Interface.基于聚类核强化学习的脑机接口三杆判别任务中的卡尔曼滤波器。
Annu Int Conf IEEE Eng Med Biol Soc. 2022 Jul;2022:690-693. doi: 10.1109/EMBC48229.2022.9871669.
9
A Kernel Reinforcement Learning Decoding Framework Integrating Neural and Feedback Signals for Brain Control.一种整合神经和反馈信号的核强化学习解码框架,用于脑控。
Annu Int Conf IEEE Eng Med Biol Soc. 2023 Jul;2023:1-4. doi: 10.1109/EMBC40787.2023.10340203.
10
Kernel Temporal Difference based Reinforcement Learning for Brain Machine Interfaces.基于核时差分的脑机接口强化学习。
Annu Int Conf IEEE Eng Med Biol Soc. 2021 Nov;2021:6721-6724. doi: 10.1109/EMBC46164.2021.9631086.