Imperial College London.
Friedrich-Alexander University Erlangen-Nuremberg.
J Cogn Neurosci. 2022 Feb 1;34(3):411-424. doi: 10.1162/jocn_a_01811.
Speech and music are spectrotemporally complex acoustic signals that are highly relevant for humans. Both contain a temporal fine structure that is encoded in the neural responses of subcortical and cortical processing centers. The subcortical response to the temporal fine structure of speech has recently been shown to be modulated by selective attention to one of two competing voices. Music similarly often consists of several simultaneous melodic lines, and a listener can selectively attend to a particular one at a time. However, the neural mechanisms that enable such selective attention remain largely enigmatic, not least since most investigations to date have focused on short and simplified musical stimuli. Here, we studied the neural encoding of classical musical pieces in human volunteers, using scalp EEG recordings. We presented volunteers with continuous musical pieces composed of one or two instruments. In the latter case, the participants were asked to selectively attend to one of the two competing instruments and to perform a vibrato identification task. We used linear encoding and decoding models to relate the recorded EEG activity to the stimulus waveform. We show that we can measure neural responses to the temporal fine structure of melodic lines played by one single instrument, at the population level as well as for most individual participants. The neural response peaks at a latency of 7.6 msec and is not measurable past 15 msec. When analyzing the neural responses to the temporal fine structure elicited by competing instruments, we found no evidence of attentional modulation. We observed, however, that low-frequency neural activity exhibited a modulation consistent with the behavioral task at latencies from 100 to 160 msec, in a similar manner to the attentional modulation observed in continuous speech (N100). Our results show that, much like speech, the temporal fine structure of music is tracked by neural activity. In contrast to speech, however, this response appears unaffected by selective attention in the context of our experiment.
言语和音乐是具有spectrotemporally 复杂的声学信号,对人类具有高度的相关性。两者都包含一个在亚皮质和皮质处理中心的神经反应中编码的时间精细结构。最近,已经表明对两个竞争声音之一的选择性注意会调制言语的时间精细结构的亚皮质反应。音乐同样经常由几个同时的旋律线组成,听众可以一次选择性地关注一个特定的旋律线。然而,使这种选择性注意成为可能的神经机制在很大程度上仍然是个谜,尤其是因为迄今为止大多数调查都集中在简短和简化的音乐刺激上。在这里,我们使用头皮 EEG 记录研究了人类志愿者对古典音乐作品的神经编码。我们向志愿者展示了由一个或两个乐器组成的连续音乐作品。在后一种情况下,要求参与者选择性地关注两个竞争乐器之一,并执行颤音识别任务。我们使用线性编码和解码模型将记录的 EEG 活动与刺激波形相关联。我们表明,我们可以在群体水平以及大多数个体参与者中测量到由单个乐器演奏的旋律线的时间精细结构的神经反应。神经反应在 7.6 msec 处达到峰值,在 15 msec 后不可测量。当分析由竞争乐器引起的时间精细结构的神经反应时,我们没有发现注意力调制的证据。然而,我们观察到,低频神经活动表现出与行为任务一致的调制,潜伏期从 100 到 160 msec,类似于连续言语中观察到的注意力调制(N100)。我们的结果表明,就像言语一样,音乐的时间精细结构也由神经活动跟踪。然而,与言语不同的是,在我们的实验中,这种反应似乎不受选择性注意的影响。