Department of Cognitive Neuroscience, Faculty of Psychology and Neuroscience, Maastricht University, P.O. Box 616, 6200MD Maastricht, The Netherlands; Maastricht Brain Imaging Centre, 6200MD Maastricht, The Netherlands.
Department of Cognitive Neuroscience, Faculty of Psychology and Neuroscience, Maastricht University, P.O. Box 616, 6200MD Maastricht, The Netherlands; Maastricht Brain Imaging Centre, 6200MD Maastricht, The Netherlands.
Neuroimage. 2021 Mar;228:117670. doi: 10.1016/j.neuroimage.2020.117670. Epub 2020 Dec 24.
Selective attention is essential for the processing of multi-speaker auditory scenes because they require the perceptual segregation of the relevant speech ("target") from irrelevant speech ("distractors"). For simple sounds, it has been suggested that the processing of multiple distractor sounds depends on bottom-up factors affecting task performance. However, it remains unclear whether such dependency applies to naturalistic multi-speaker auditory scenes. In this study, we tested the hypothesis that increased perceptual demand (the processing requirement posed by the scene to separate the target speech) reduces the cortical processing of distractor speech thus decreasing their perceptual segregation. Human participants were presented with auditory scenes including three speakers and asked to selectively attend to one speaker while their EEG was acquired. The perceptual demand of this selective listening task was varied by introducing an auditory cue (interaural time differences, ITDs) for segregating the target from the distractor speakers, while acoustic differences between the distractors were matched in ITD and loudness. We obtained a quantitative measure of the cortical segregation of distractor speakers by assessing the difference in how accurately speech-envelope following EEG responses could be predicted by models of averaged distractor speech versus models of individual distractor speech. In agreement with our hypothesis, results show that interaural segregation cues led to improved behavioral word-recognition performance and stronger cortical segregation of the distractor speakers. The neural effect was strongest in the δ-band and at early delays (0 - 200 ms). Our results indicate that during low perceptual demand, the human cortex represents individual distractor speech signals as more segregated. This suggests that, in addition to purely acoustical properties, the cortical processing of distractor speakers depends on factors like perceptual demand.
选择性注意对于多说话人听觉场景的处理至关重要,因为它们需要将相关语音(“目标”)与不相关语音(“干扰”)进行感知分离。对于简单的声音,有人认为,多个干扰声音的处理取决于影响任务表现的自下而上的因素。然而,目前尚不清楚这种依赖性是否适用于自然多说话人听觉场景。在这项研究中,我们检验了这样一种假设,即增加感知需求(场景对分离目标语音提出的处理要求)会降低对干扰语音的皮质处理,从而减少其感知分离。研究人员向人类参与者呈现了包括三个说话者的听觉场景,并要求他们在获取 EEG 的同时选择性地关注一个说话者。通过引入听觉线索(耳间时间差,ITD)将目标与干扰说话者分开,从而改变选择性聆听任务的感知需求,同时在 ITD 和响度上匹配干扰者之间的声学差异。我们通过评估平均干扰语音模型与个别干扰语音模型对语音包络跟随 EEG 响应的预测精度差异,获得了干扰语音皮质分离的定量测量值。与我们的假设一致,结果表明,耳间分离线索导致行为上单词识别性能的提高和干扰者的皮质分离增强。该神经效应在 δ 波段和早期延迟(0-200ms)最强。我们的研究结果表明,在低感知需求下,人类大脑会将个别干扰语音信号表示为更具分离性。这表明,除了纯粹的声学特性外,干扰语音的皮质处理还取决于感知需求等因素。