Love Scott A, Pollick Frank E, Latinus Marianne
School of Psychology, University of Glasgow, 58 Hillhead Street, Glasgow G12 8QB, UK.
Seeing Perceiving. 2011;24(4):351-67. doi: 10.1163/187847511X584452.
Perception of faces and voices plays a prominent role in human social interaction, making multisensory integration of cross-modal speech a topic of great interest in cognitive neuroscience. How to define potential sites of multisensory integration using functional magnetic resonance imaging (fMRI) is currently under debate, with three statistical criteria frequently used (e.g., super-additive, max and mean criteria). In the present fMRI study, 20 participants were scanned in a block design under three stimulus conditions: dynamic unimodal face, unimodal voice and bimodal face-voice. Using this single dataset, we examine all these statistical criteria in an attempt to define loci of face-voice integration. While the super-additive and mean criteria essentially revealed regions in which one of the unimodal responses was a deactivation, the max criterion appeared stringent and only highlighted the left hippocampus as a potential site of face- voice integration. Psychophysiological interaction analysis showed that connectivity between occipital and temporal cortices increased during bimodal compared to unimodal conditions. We concluded that, when investigating multisensory integration with fMRI, all these criteria should be used in conjunction with manipulation of stimulus signal-to-noise ratio and/or cross-modal congruency.
对面孔和声音的感知在人类社交互动中起着重要作用,这使得跨模态语音的多感官整合成为认知神经科学中一个备受关注的话题。如何使用功能磁共振成像(fMRI)来定义多感官整合的潜在部位目前仍存在争议,常用的有三种统计标准(例如,超加性、最大值和平均值标准)。在本fMRI研究中,20名参与者在组块设计下接受了三种刺激条件的扫描:动态单模态面孔、单模态声音和双模态面孔-声音。利用这个单一数据集,我们检验了所有这些统计标准,试图确定面孔-声音整合的位点。虽然超加性和平均值标准基本上揭示了单模态反应之一为失活的区域,但最大值标准显得更为严格,仅突出了左侧海马体作为面孔-声音整合的一个潜在部位。心理生理交互作用分析表明,与单模态条件相比,双模态条件下枕叶和颞叶皮质之间的连通性增加。我们得出结论,在使用fMRI研究多感官整合时,所有这些标准都应与刺激信噪比和/或跨模态一致性的操作结合使用。