Bestelmeyer Patricia E G, Maurage Pierre, Rouger Julien, Latinus Marianne, Belin Pascal
School of Psychology, Bangor University, Bangor, Gwynedd LL57 2AS, United Kingdom,
Cognitive Neurosciences and Clinical Psychology Research Units, Department of Psychology, Catholic University of Louvain, 1348 Louvain-la-Neuve, Belgium.
J Neurosci. 2014 Jun 11;34(24):8098-105. doi: 10.1523/JNEUROSCI.4820-13.2014.
The human voice carries speech as well as important nonlinguistic signals that influence our social interactions. Among these cues that impact our behavior and communication with other people is the perceived emotional state of the speaker. A theoretical framework for the neural processing stages of emotional prosody has suggested that auditory emotion is perceived in multiple steps (Schirmer and Kotz, 2006) involving low-level auditory analysis and integration of the acoustic information followed by higher-level cognition. Empirical evidence for this multistep processing chain, however, is still sparse. We examined this question using functional magnetic resonance imaging and a continuous carry-over design (Aguirre, 2007) to measure brain activity while volunteers listened to non-speech-affective vocalizations morphed on a continuum between anger and fear. Analyses dissociated neuronal adaptation effects induced by similarity in perceived emotional content between consecutive stimuli from those induced by their acoustic similarity. We found that bilateral voice-sensitive auditory regions as well as right amygdala coded the physical difference between consecutive stimuli. In contrast, activity in bilateral anterior insulae, medial superior frontal cortex, precuneus, and subcortical regions such as bilateral hippocampi depended predominantly on the perceptual difference between morphs. Our results suggest that the processing of vocal affect recognition is a multistep process involving largely distinct neural networks. Amygdala and auditory areas predominantly code emotion-related acoustic information while more anterior insular and prefrontal regions respond to the abstract, cognitive representation of vocal affect.
人类的声音不仅传递言语,还携带重要的非语言信号,这些信号会影响我们的社交互动。在这些影响我们与他人行为及交流的线索中,说话者的感知情绪状态是其中之一。一个关于情感韵律神经处理阶段的理论框架表明,听觉情绪是通过多个步骤被感知的(施尔默和科茨,2006年),包括低级听觉分析和声学信息整合,随后是高级认知。然而,这一多步骤处理链的实证证据仍然稀少。我们使用功能磁共振成像和连续遗留设计(阿吉雷,2007年)来研究这个问题,以测量志愿者在听愤怒和恐惧之间连续渐变的非言语情感发声时的大脑活动。分析将连续刺激之间感知到的情感内容相似性所引发的神经元适应效应,与由声学相似性所引发的效应区分开来。我们发现,双侧对声音敏感的听觉区域以及右侧杏仁核编码了连续刺激之间的物理差异。相比之下,双侧前脑岛、内侧额上回、楔前叶以及双侧海马体等皮质下区域的活动主要取决于变体之间的感知差异。我们的结果表明,声音情感识别的处理是一个多步骤过程,涉及很大程度上不同的神经网络。杏仁核和听觉区域主要编码与情感相关的声学信息,而更靠前的脑岛和前额叶区域则对声音情感的抽象认知表征做出反应。