Center of Cognitive Neuroscience, University of Salzburg, Salzburg, Austria.
Department of Psychology, University of Salzburg, Salzburg, Austria.
Eur J Neurosci. 2022 Jun;55(11-12):3288-3302. doi: 10.1111/ejn.14912. Epub 2020 Aug 7.
Making sense of a poor auditory signal can pose a challenge. Previous attempts to quantify speech intelligibility in neural terms have usually focused on one of two measures, namely low-frequency speech-brain synchronization or alpha power modulations. However, reports have been mixed concerning the modulation of these measures, an issue aggravated by the fact that they have normally been studied separately. We present two MEG studies analyzing both measures. In study 1, participants listened to unimodal auditory speech with three different levels of degradation (original, 7-channel and 3-channel vocoding). Intelligibility declined with declining clarity, but speech was still intelligible to some extent even for the lowest clarity level (3-channel vocoding). Low-frequency (1-7 Hz) speech tracking suggested a U-shaped relationship with strongest effects for the medium-degraded speech (7-channel) in bilateral auditory and left frontal regions. To follow up on this finding, we implemented three additional vocoding levels (5-channel, 2-channel and 1-channel) in a second MEG study. Using this wider range of degradation, the speech-brain synchronization showed a similar pattern as in study 1, but further showed that when speech becomes unintelligible, synchronization declines again. The relationship differed for alpha power, which continued to decrease across vocoding levels reaching a floor effect for 5-channel vocoding. Predicting subjective intelligibility based on models either combining both measures or each measure alone showed superiority of the combined model. Our findings underline that speech tracking and alpha power are modified differently by the degree of degradation of continuous speech but together contribute to the subjective speech understanding.
理解质量差的听觉信号可能具有挑战性。先前在神经学方面量化言语可懂度的尝试通常集中在两个测量方法之一,即低频语音-脑同步或α 功率调制。然而,关于这些测量方法的调制的报告结果不一,这一问题因它们通常分别进行研究而加剧。我们提出了两项 MEG 研究来分析这两个测量方法。在研究 1 中,参与者收听具有三种不同退化水平的单模态听觉语音(原始、7 通道和 3 通道声码化)。可懂度随清晰度下降而下降,但即使在最低清晰度水平(3 通道声码化)下,语音仍具有一定的可懂度。低频(1-7 Hz)语音跟踪表明,在双边听觉和左额区,具有最强影响的是中退化语音(7 通道)的 U 形关系。为了跟进这一发现,我们在第二项 MEG 研究中实施了三个额外的声码化水平(5 通道、2 通道和 1 通道)。使用更广泛的退化范围,语音-脑同步显示出与研究 1 相似的模式,但进一步表明,当语音变得不可懂时,同步再次下降。α 功率的关系不同,其随着声码化水平的降低而继续降低,在 5 通道声码化时达到地板效应。基于结合这两个测量方法或每个测量方法单独的模型来预测主观可懂度,显示出组合模型的优越性。我们的发现强调了语音跟踪和α 功率在连续语音的退化程度上受到不同程度的影响,但它们共同有助于主观的语音理解。