Lachs Lorin, Pisoni David B
Department of Psychology, California State University, Fresno, CA, USA.
J Exp Psychol Hum Percept Perform. 2004 Apr;30(2):378-96. doi: 10.1037/0096-1523.30.2.378.
In a cross-modal matching task, participants were asked to match visual and auditory displays of speech based on the identity of the speaker. The present investigation used this task with acoustically transformed speech to examine the properties of sound that can convey cross-modal information. Word recognition performance was also measured under the same transformations. The authors found that cross-modal matching was only possible under transformations that preserved the relative spectral and temporal patterns of formant frequencies. In addition, cross-modal matching was only possible under the same conditions that yielded robust word recognition performance. The results are consistent with the hypothesis that acoustic and optical displays of speech simultaneously carry articulatory information about both the underlying linguistic message and indexical properties of the talker.
在一个跨模态匹配任务中,参与者被要求根据说话者的身份匹配语音的视觉和听觉展示。本研究使用这个任务以及经过声学转换的语音来检验能够传达跨模态信息的声音特性。在相同的转换条件下还测量了单词识别表现。作者发现,只有在保留共振峰频率的相对频谱和时间模式的转换条件下,跨模态匹配才有可能实现。此外,只有在产生稳健单词识别表现的相同条件下,跨模态匹配才有可能实现。这些结果与以下假设一致,即语音的声学和光学展示同时携带关于潜在语言信息和说话者索引属性的发音信息。