Bujok Ronny, Peeters David, Meyer Antje S, Bosker Hans Rutger
Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands.
International Max Planck Research School for Language Sciences, MPI for Psycholinguistics, Max Planck Society, Nijmegen, The Netherlands.
Atten Percept Psychophys. 2025 May 20. doi: 10.3758/s13414-025-03088-5.
Speech is inherently variable, requiring listeners to apply adaptation mechanisms to deal with the variability. A proposed perceptual adaptation mechanism is recalibration, whereby listeners learn to adjust cognitive representations of speech sounds based on disambiguating contextual information. Most studies on the role of recalibration in speech perception have focused on variability in particular speech segments (e.g., consonants/vowels), and speech has mostly been studied with a focus on talking heads. However, speech is often accompanied by visual bodily signals like hand gestures, and is thus multimodal. Moreover, variability in speech extends beyond segmental aspects alone and also affects prosodic aspects, like lexical stress. We currently do not understand well how listeners adjust their representations of lexical stress patterns to different speakers. In four experiments, we investigated recalibration of lexical stress perception, driven by lexico-orthographical information (Experiment 1) and by manual beat gestures (Experiments 2-4). Across experiments, we observed that these two types of disambiguating information (presented in an audiovisual exposure phase) led listeners to adjust their representations of lexical stress, with lasting consequences for subsequent spoken word recognition (in an audio-only test phase). However, evidence for generalization of this recalibration to new words was only found in the third experiment, suggesting that generalization may be limited. These results highlight that recalibration is a plausible mechanism for suprasegmental speech adaption in everyday communication and show that even the timing of simple hand gestures can have a lasting effect on auditory speech perception.
言语本质上是可变的,这就要求听众运用适应机制来应对这种变异性。一种提出的感知适应机制是重新校准,即听众学会根据消除歧义的上下文信息来调整语音的认知表征。大多数关于重新校准在言语感知中作用的研究都集中在特定言语片段(如辅音/元音)的变异性上,并且言语大多是通过关注会说话的头像来进行研究的。然而,言语通常会伴随着诸如手势等视觉身体信号,因此是多模态的。此外,言语的变异性不仅限于片段层面,还会影响韵律方面,如词汇重音。我们目前还不太清楚听众如何根据不同的说话者来调整他们对词汇重音模式的表征。在四项实验中,我们研究了由词汇正字法信息(实验1)和手动节拍手势(实验2 - 4)驱动的词汇重音感知的重新校准。在所有实验中,我们观察到这两种消除歧义的信息(在视听暴露阶段呈现)会使听众调整他们对词汇重音的表征,对随后的口语单词识别(在仅音频测试阶段)产生持久影响。然而,这种重新校准对新单词的泛化证据仅在第三个实验中被发现,这表明泛化可能是有限的。这些结果突出表明,重新校准是日常交流中用于超音段言语适应的一种合理机制,并表明即使是简单手势的时机也会对听觉言语感知产生持久影响。