Lachs Lorin, Pisoni David B
Department of Psychology California State University, Fresno.
Ecol Psychol. 2004;16(3):159-187. doi: 10.1207/s15326969eco1603_1.
Four experiments examined the nature of multisensory speech information. In Experiment 1, participants were asked to match heard voices with dynamic visual-alone video clips of speakers' articulating faces. This cross-modal matching task was used to examine whether vocal source matching can be accomplished across sensory modalities. The results showed that observers could match speaking faces and voices, indicating that information about the speaker was available for cross-modal comparisons. In a series of follow-up experiments, several stimulus manipulations were used to determine some of the critical acoustic and optic patterns necessary for specifying cross-modal source information. The results showed that cross-modal source information was not available in static visual displays of faces and was not contingent on a prominent acoustic cue to vocal identity (f0). Furthermore, cross-modal matching was not possible when the acoustic signal was temporally reversed.
四项实验研究了多感官语音信息的本质。在实验1中,参与者被要求将听到的声音与说话者面部发音的动态纯视觉视频片段进行匹配。这个跨模态匹配任务用于检验语音源匹配是否可以在不同感官模态之间完成。结果表明,观察者能够匹配说话的面部和声音,这表明关于说话者的信息可用于跨模态比较。在一系列后续实验中,采用了几种刺激操作来确定指定跨模态源信息所需的一些关键声学和光学模式。结果表明,跨模态源信息在面部的静态视觉显示中不可用,并且不取决于对语音身份的突出声学线索(基频)。此外,当声学信号在时间上反转时,跨模态匹配是不可能的。