Radboud University.
Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands.
J Cogn Neurosci. 2018 Aug;30(8):1086-1097. doi: 10.1162/jocn_a_01301. Epub 2018 Jun 19.
Previous work revealed that visual semantic information conveyed by gestures can enhance degraded speech comprehension, but the mechanisms underlying these integration processes under adverse listening conditions remain poorly understood. We used MEG to investigate how oscillatory dynamics support speech-gesture integration when integration load is manipulated by auditory (e.g., speech degradation) and visual semantic (e.g., gesture congruency) factors. Participants were presented with videos of an actress uttering an action verb in clear or degraded speech, accompanied by a matching (mixing gesture + "mixing") or mismatching (drinking gesture + "walking") gesture. In clear speech, alpha/beta power was more suppressed in the left inferior frontal gyrus and motor and visual cortices when integration load increased in response to mismatching versus matching gestures. In degraded speech, beta power was less suppressed over posterior STS and medial temporal lobe for mismatching compared with matching gestures, showing that integration load was lowest when speech was degraded and mismatching gestures could not be integrated and disambiguate the degraded signal. Our results thus provide novel insights on how low-frequency oscillatory modulations in different parts of the cortex support the semantic audiovisual integration of gestures in clear and degraded speech: When speech is clear, the left inferior frontal gyrus and motor and visual cortices engage because higher-level semantic information increases semantic integration load. When speech is degraded, posterior STS/middle temporal gyrus and medial temporal lobe are less engaged because integration load is lowest when visual semantic information does not aid lexical retrieval and speech and gestures cannot be integrated.
先前的工作表明,手势所传达的视觉语义信息可以增强受损语音的理解,但在不利的听力条件下,这些整合过程的机制仍未被充分理解。我们使用 MEG 研究了在听觉(例如语音退化)和视觉语义(例如手势一致性)因素操纵整合负载的情况下,振荡动力学如何支持语音-手势整合。参与者观看女演员用清晰或受损的语音说出一个动作动词的视频,同时伴有匹配(混合手势+“混合”)或不匹配(喝水手势+“行走”)的手势。在清晰的语音中,当整合负载响应于不匹配与匹配的手势而增加时,左额下回、运动和视觉皮层的α/β功率抑制更大。在受损的语音中,与匹配的手势相比,不匹配的手势在 STS 后和内侧颞叶的β功率抑制较小,表明当语音退化且不匹配的手势无法整合并消除退化信号时,整合负载最低。因此,我们的研究结果提供了关于低频振荡调制如何支持清晰和受损语音中手势的语义视听整合的新见解:当语音清晰时,左额下回、运动和视觉皮层参与,因为高级语义信息增加了语义整合负载。当语音退化时,STS 后区/颞中回和内侧颞叶参与较少,因为当视觉语义信息无助于词汇检索并且语音和手势无法整合时,整合负载最低。