Sohoglu Ediz, Davis Matthew H
School of Psychology, University of Sussex, Brighton, United Kingdom.
MRC Cognition and Brain Sciences Unit, Cambridge, United Kingdom.
Elife. 2020 Nov 4;9:e58077. doi: 10.7554/eLife.58077.
Human speech perception can be described as Bayesian perceptual inference but how are these Bayesian computations instantiated neurally? We used magnetoencephalographic recordings of brain responses to degraded spoken words and experimentally manipulated signal quality and prior knowledge. We first demonstrate that spectrotemporal modulations in speech are more strongly represented in neural responses than alternative speech representations (e.g. spectrogram or articulatory features). Critically, we found an interaction between speech signal quality and expectations from prior written text on the quality of neural representations; increased signal quality enhanced neural representations of speech that mismatched with prior expectations, but led to greater suppression of speech that matched prior expectations. This interaction is a unique neural signature of prediction error computations and is apparent in neural responses within 100 ms of speech input. Our findings contribute to the detailed specification of a computational model of speech perception based on predictive coding frameworks.
人类语音感知可被描述为贝叶斯感知推理,但这些贝叶斯计算是如何在神经层面实现的呢?我们利用脑磁图记录大脑对 degraded 口语单词的反应,并通过实验操纵信号质量和先验知识。我们首先证明,语音中的频谱时间调制在神经反应中比其他语音表征(如频谱图或发音特征)得到更强的表征。至关重要的是,我们发现语音信号质量与来自先前书面文本对神经表征质量的期望之间存在相互作用;信号质量的提高增强了与先前期望不匹配的语音的神经表征,但导致与先前期望匹配的语音受到更大程度的抑制。这种相互作用是预测误差计算的独特神经特征,并且在语音输入后100毫秒内的神经反应中很明显。我们的研究结果有助于基于预测编码框架详细说明语音感知的计算模型。