Veillette John P, Rosen Jacob, Margoliash Daniel, Nusbaum Howard C
Departments of Psychology, University of Chicago Chicago, Illinois 60637
Departments of Psychology, University of Chicago Chicago, Illinois 60637.
J Neurosci. 2025 May 21;45(21):e2294242025. doi: 10.1523/JNEUROSCI.2294-24.2025.
To learn complex motor skills, an organism must be able to assign sensory feedback events to the actions that caused them. This matching problem would be simple if motor neuron output led to sensory feedback with a fixed, predictable lag. However, nonlinear dynamics in the brain and the body's periphery can decouple the timing of critical events from that of the motor output which caused them. During human speech production, for example, phonation from the glottis (a sound source for speech) begins suddenly when subglottal pressure and laryngeal tension cross a sharp threshold (i.e., a bifurcation). Only if the brain can predict the timing of these discrete peripheral events resulting from motor output, then, would it be possible to match sensory feedback to movements based on temporal coherence. We show that event onsets in the (male and female) human glottal waveform, measured using electroglottography, are reflected in the electroencephalogram during speech production, leading up to the time of the event itself. Conversely, glottal event times can be decoded from the electroencephalogram. After prolonged exposure to delayed auditory feedback, subjects recalibrate their behavioral threshold for detecting temporal auditory-motor mismatches and decoded event times decouple from actual movements. This suggests decoding performance is driven by plastic predictions of peripheral timing, providing a missing component for hindsight credit assignment, in which specific feedback events are associated with the neural activity that gave rise to movements. We discuss parallel findings from the birdsong system suggesting that results may generalize across vocal learning species.
为了学习复杂的运动技能,生物体必须能够将感觉反馈事件与引发这些事件的动作联系起来。如果运动神经元输出导致感觉反馈具有固定、可预测的延迟,那么这个匹配问题将会很简单。然而,大脑和身体外周的非线性动力学能够使关键事件的时间与引发这些事件的运动输出的时间解耦。例如,在人类言语产生过程中,当声门下压力和喉部张力越过一个急剧的阈值(即一个分岔点)时,声门(言语的一个声源)的发声会突然开始。只有当大脑能够预测由运动输出产生的这些离散外周事件的时间时,才有可能基于时间连贯性将感觉反馈与动作相匹配。我们表明,使用电声门图测量的(男性和女性)人类声门波形中的事件起始,在言语产生过程中直至事件本身发生的时刻,都会反映在脑电图中。相反,声门事件时间可以从脑电图中解码出来。在长时间暴露于延迟听觉反馈后,受试者会重新校准他们检测时间性听觉 - 运动不匹配的行为阈值,并且解码出的事件时间与实际动作解耦。这表明解码性能是由对外周时间的可塑性预测驱动的,为事后归因提供了一个缺失的组成部分,在事后归因中,特定的反馈事件与引发动作的神经活动相关联。我们讨论了来自鸟鸣系统的平行发现,表明这些结果可能适用于所有声乐学习物种。