Lachs Lorin, Pisoni David B
Department of Psychology, 5310 North Campus Drive, California State University, Fresno, California 93740, USA.
J Acoust Soc Am. 2004 Jul;116(1):507-18. doi: 10.1121/1.1757454.
Information about the acoustic properties of a talker's voice is available in optical displays of speech, and vice versa, as evidenced by perceivers' ability to match faces and voices based on vocal identity. The present investigation used point-light displays (PLDs) of visual speech and sinewave replicas of auditory speech in a cross-modal matching task to assess perceivers' ability to match faces and voices under conditions when only isolated kinematic information about vocal tract articulation was available. These stimuli were also used in a word recognition experiment under auditory-alone and audiovisual conditions. The results showed that isolated kinematic displays provide enough information to match the source of an utterance across sensory modalities. Furthermore, isolated kinematic displays can be integrated to yield better word recognition performance under audiovisual conditions than under auditory-alone conditions. The results are discussed in terms of their implications for describing the nature of speech information and current theories of speech perception and spoken word recognition.
说话者声音的声学特性信息可在语音的光学显示中获取,反之亦然,这一点已通过感知者基于声音特征匹配面孔和声音的能力得到证明。本研究在跨模态匹配任务中使用了视觉语音的点光显示(PLD)和听觉语音的正弦波复制品,以评估在仅提供有关声道发音的孤立运动学信息的条件下,感知者匹配面孔和声音的能力。这些刺激还用于仅听觉和视听条件下的单词识别实验。结果表明,孤立的运动学显示提供了足够的信息来跨感官模态匹配话语的来源。此外,在视听条件下,孤立的运动学显示可以整合起来,从而产生比仅听觉条件下更好的单词识别性能。我们将根据这些结果对描述语音信息的本质以及当前语音感知和口语单词识别理论的意义进行讨论。