• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

量化基于奖励的运动学习中探索的陷阱,以及如何避免这些陷阱。

Pitfalls in quantifying exploration in reward-based motor learning and how to avoid them.

机构信息

Department of Human Movement Sciences, Vrije Universiteit Amsterdam, Amsterdam, The Netherlands.

出版信息

Biol Cybern. 2021 Aug;115(4):365-382. doi: 10.1007/s00422-021-00884-8. Epub 2021 Aug 2.

DOI:10.1007/s00422-021-00884-8
PMID:34341885
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8382626/
Abstract

When learning a movement based on binary success information, one is more variable following failure than following success. Theoretically, the additional variability post-failure might reflect exploration of possibilities to obtain success. When average behavior is changing (as in learning), variability can be estimated from differences between subsequent movements. Can one estimate exploration reliably from such trial-to-trial changes when studying reward-based motor learning? To answer this question, we tried to reconstruct the exploration underlying learning as described by four existing reward-based motor learning models. We simulated learning for various learner and task characteristics. If we simply determined the additional change post-failure, estimates of exploration were sensitive to learner and task characteristics. We identified two pitfalls in quantifying exploration based on trial-to-trial changes. Firstly, performance-dependent feedback can cause correlated samples of motor noise and exploration on successful trials, which biases exploration estimates. Secondly, the trial relative to which trial-to-trial change is calculated may also contain exploration, which causes underestimation. As a solution, we developed the additional trial-to-trial change (ATTC) method. By moving the reference trial one trial back and subtracting trial-to-trial changes following specific sequences of trial outcomes, exploration can be estimated reliably for the three models that explore based on the outcome of only the previous trial. Since ATTC estimates are based on a selection of trial sequences, this method requires many trials. In conclusion, if exploration is a binary function of previous trial outcome, the ATTC method allows for a model-free quantification of exploration.

摘要

当基于二进制成功信息学习一个动作时,失败后比成功后更容易发生变化。从理论上讲,失败后额外的可变性可能反映了获得成功的可能性探索。当平均行为发生变化时(例如在学习中),可以从后续运动之间的差异来估计可变性。在研究基于奖励的运动学习时,是否可以从这种逐次试验的变化中可靠地估计探索?为了回答这个问题,我们试图根据四个现有的基于奖励的运动学习模型来重建学习中的探索。我们模拟了各种学习者和任务特征的学习。如果我们只是确定失败后的额外变化,那么探索的估计值就会受到学习者和任务特征的影响。我们确定了在基于逐次试验的变化量化探索中的两个陷阱。首先,性能相关的反馈会导致成功试验中运动噪声和探索的相关样本,从而产生探索的偏差估计。其次,相对于计算逐次试验变化的试验本身可能也包含探索,从而导致低估。作为一种解决方案,我们开发了额外的逐次试验变化(ATTC)方法。通过将参考试验向后移动一个试验,并减去以下特定试验结果序列的逐次试验变化,可以可靠地估计基于仅前一个试验结果的三个模型的探索。由于 ATTC 估计值基于试验序列的选择,因此该方法需要许多试验。总之,如果探索是前一个试验结果的二进制函数,那么 ATTC 方法允许对探索进行无模型的量化。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/41bd/8382626/b744ceec8af9/422_2021_884_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/41bd/8382626/4dcdd6c0c114/422_2021_884_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/41bd/8382626/97fc1ff06d41/422_2021_884_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/41bd/8382626/f7d9a00eea9b/422_2021_884_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/41bd/8382626/348bfec7201f/422_2021_884_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/41bd/8382626/8522b7fbfc76/422_2021_884_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/41bd/8382626/c19756baa21a/422_2021_884_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/41bd/8382626/e9d2d8ab15c2/422_2021_884_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/41bd/8382626/cd19a102ef07/422_2021_884_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/41bd/8382626/b744ceec8af9/422_2021_884_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/41bd/8382626/4dcdd6c0c114/422_2021_884_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/41bd/8382626/97fc1ff06d41/422_2021_884_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/41bd/8382626/f7d9a00eea9b/422_2021_884_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/41bd/8382626/348bfec7201f/422_2021_884_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/41bd/8382626/8522b7fbfc76/422_2021_884_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/41bd/8382626/c19756baa21a/422_2021_884_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/41bd/8382626/e9d2d8ab15c2/422_2021_884_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/41bd/8382626/cd19a102ef07/422_2021_884_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/41bd/8382626/b744ceec8af9/422_2021_884_Fig9_HTML.jpg

相似文献

1
Pitfalls in quantifying exploration in reward-based motor learning and how to avoid them.量化基于奖励的运动学习中探索的陷阱,以及如何避免这些陷阱。
Biol Cybern. 2021 Aug;115(4):365-382. doi: 10.1007/s00422-021-00884-8. Epub 2021 Aug 2.
2
Quantifying exploration in reward-based motor learning.量化基于奖励的运动学习中的探索。
PLoS One. 2020 Apr 2;15(4):e0226789. doi: 10.1371/journal.pone.0226789. eCollection 2020.
3
Alterations in the amplitude and burst rate of beta oscillations impair reward-dependent motor learning in anxiety.β 振荡幅度和爆发率的改变会损害焦虑症患者的奖励依赖性运动学习。
Elife. 2020 May 19;9:e50654. doi: 10.7554/eLife.50654.
4
Somatic and Reinforcement-Based Plasticity in the Initial Stages of Human Motor Learning.人类运动学习初期基于躯体和强化的可塑性
J Neurosci. 2016 Nov 16;36(46):11682-11692. doi: 10.1523/JNEUROSCI.1767-16.2016.
5
Clustering analysis of movement kinematics in reinforcement learning.强化学习中运动运动学的聚类分析。
J Neurophysiol. 2022 Feb 1;127(2):341-353. doi: 10.1152/jn.00229.2021. Epub 2021 Dec 22.
6
Interactions between motor exploration and reinforcement learning.运动探索与强化学习的相互作用。
J Neurophysiol. 2019 Aug 1;122(2):797-808. doi: 10.1152/jn.00390.2018. Epub 2019 Jun 26.
7
Task-relevant and task-irrelevant variability causally shape error-based motor learning.任务相关和任务无关的可变性因果塑造基于错误的运动学习。
Neural Netw. 2021 Oct;142:583-596. doi: 10.1016/j.neunet.2021.07.015. Epub 2021 Jul 21.
8
Failure induces task-irrelevant exploration during a stencil task.失败会在模板任务中诱导出与任务无关的探索。
Exp Brain Res. 2023 Feb;241(2):677-686. doi: 10.1007/s00221-023-06548-2. Epub 2023 Jan 20.
9
When money is not enough: awareness, success, and variability in motor learning.资金不足时:运动学习中的意识、成功与变异性。
PLoS One. 2014 Jan 28;9(1):e86580. doi: 10.1371/journal.pone.0086580. eCollection 2014.
10
Contribution of explicit processes to reinforcement-based motor learning.显性过程对基于强化的运动学习的贡献。
J Neurophysiol. 2018 Jun 1;119(6):2241-2255. doi: 10.1152/jn.00901.2017. Epub 2018 Mar 14.

引用本文的文献

1
The sign of exploration during reward-based motor learning is not independent from trial to trial.基于奖励的运动学习过程中的探索信号在每次试验之间并非相互独立。
Exp Brain Res. 2025 Apr 15;243(5):117. doi: 10.1007/s00221-025-07074-z.
2
The effect of transcranial random noise stimulation on the movement time and components of noise, co-variation, and tolerance in a perceptual-motor task.经颅随机噪声刺激对感知运动任务中运动时间以及噪声成分、协变和耐受性的影响。
Sci Rep. 2025 Feb 3;15(1):4083. doi: 10.1038/s41598-025-88396-4.
3
Roles and interplay of reinforcement-based and error-based processes during reaching and gait in neurotypical adults and individuals with Parkinson's disease.

本文引用的文献

1
Quantifying exploration in reward-based motor learning.量化基于奖励的运动学习中的探索。
PLoS One. 2020 Apr 2;15(4):e0226789. doi: 10.1371/journal.pone.0226789. eCollection 2020.
2
Adaptive Regulation of Motor Variability.运动可变性的自适应调节。
Curr Biol. 2019 Nov 4;29(21):3551-3562.e7. doi: 10.1016/j.cub.2019.08.052. Epub 2019 Oct 17.
3
Interactions between motor exploration and reinforcement learning.运动探索与强化学习的相互作用。
在神经正常的成年人和帕金森病患者的伸手和步态中,基于强化和基于错误的过程的作用和相互作用。
PLoS Comput Biol. 2024 Oct 14;20(10):e1012474. doi: 10.1371/journal.pcbi.1012474. eCollection 2024 Oct.
4
Reinforcement Learning during Locomotion.运动过程中的强化学习。
eNeuro. 2024 Mar 15;11(3). doi: 10.1523/ENEURO.0383-23.2024. Print 2024 Mar.
5
Punishment Leads to Greater Sensorimotor Learning But Less Movement Variability Compared to Reward.与奖励相比,惩罚会导致更大的感觉运动学习,但运动变化性更小。
Neuroscience. 2024 Mar 5;540:12-26. doi: 10.1016/j.neuroscience.2024.01.004. Epub 2024 Jan 12.
6
Reinforcement-based processes actively regulate motor exploration along redundant solution manifolds.基于强化的过程沿冗余解流形主动调节运动探索。
Proc Biol Sci. 2023 Oct 25;290(2009):20231475. doi: 10.1098/rspb.2023.1475. Epub 2023 Oct 18.
7
Implicit reward-based motor learning.基于隐性奖励的运动学习。
bioRxiv. 2023 Jun 28:2023.06.27.546738. doi: 10.1101/2023.06.27.546738.
8
Parental praise and children's exploration: a virtual reality experiment.父母表扬与孩子探索:虚拟现实实验。
Sci Rep. 2022 Mar 23;12(1):4967. doi: 10.1038/s41598-022-08226-9.
J Neurophysiol. 2019 Aug 1;122(2):797-808. doi: 10.1152/jn.00390.2018. Epub 2019 Jun 26.
4
The gradient of the reinforcement landscape influences sensorimotor learning.强化景观的梯度影响感觉运动学习。
PLoS Comput Biol. 2019 Mar 4;15(3):e1006839. doi: 10.1371/journal.pcbi.1006839. eCollection 2019 Mar.
5
Neural signatures of reward and sensory error feedback processing in motor learning.运动学习中奖励和感觉错误反馈处理的神经特征。
J Neurophysiol. 2019 Apr 1;121(4):1561-1574. doi: 10.1152/jn.00792.2018. Epub 2019 Feb 27.
6
Somatosensory working memory in human reinforcement-based motor learning.人类基于强化的运动学习中的体感工作记忆。
J Neurophysiol. 2018 Dec 1;120(6):3275-3286. doi: 10.1152/jn.00442.2018. Epub 2018 Oct 24.
7
Increasing Motor Noise Impairs Reinforcement Learning in Healthy Individuals.电机噪声增大损害健康个体的强化学习能力。
eNeuro. 2018 Aug 13;5(3). doi: 10.1523/ENEURO.0050-18.2018. eCollection 2018 May-Jun.
8
Individual Differences in Motor Noise and Adaptation Rate Are Optimally Related.个体在运动噪音和适应率上的差异呈最佳相关。
eNeuro. 2018 Jul 31;5(4). doi: 10.1523/ENEURO.0170-18.2018. eCollection 2018 Jul-Aug.
9
It's Not (Only) the Mean that Matters: Variability, Noise and Exploration in Skill Learning.重要的不只是均值:技能学习中的变异性、噪声与探索
Curr Opin Behav Sci. 2018 Apr;20:183-195. doi: 10.1016/j.cobeha.2018.01.004. Epub 2018 Mar 1.
10
The relationship between reinforcement and explicit control during visuomotor adaptation.在视动适应过程中强化与显式控制之间的关系。
Sci Rep. 2018 Jun 14;8(1):9121. doi: 10.1038/s41598-018-27378-1.