Department of Hearing and Speech Sciences, Vanderbilt University Medical Center, Nashville, Tennessee, USA; Multimodal Language Department, Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands.
Department of Psychology and Human Development, Vanderbilt University, Nashville, Tennessee, USA.
Cortex. 2024 Dec;181:26-46. doi: 10.1016/j.cortex.2024.08.008. Epub 2024 Oct 22.
Language is multimodal and situated in rich visual contexts. Language is also incremental, unfolding moment-to-moment in real time, yet few studies have examined how spoken language interacts with gesture and visual context during multimodal language processing. Gesture is a rich communication cue that is integrally related to speech and often depicts concrete referents from the visual world. Using eye-tracking in an adapted visual world paradigm, we examined how participants with and without moderate-severe traumatic brain injury (TBI) use gesture to resolve temporary referential ambiguity.
Participants viewed a screen with four objects and one video. The speaker in the video produced sentences (e.g., "The girl will eat the very good sandwich"), paired with either a meaningful gesture (e.g., sandwich-holding gesture) or a meaningless grooming movement (e.g., arm scratch) at the verb "will eat." We measured participants' gaze to the target object (e.g., sandwich), a semantic competitor (e.g., apple), and two unrelated distractors (e.g., piano, guitar) during the critical window between movement onset in the gesture modality and onset of the spoken referent in speech.
Both participants with and without TBI were more likely to fixate the target when the speaker produced a gesture compared to a grooming movement; however, relative to non-injured participants, the effect was significantly attenuated in the TBI group.
We demonstrated evidence of reduced speech-gesture integration in participants with TBI relative to non-injured peers. This study advances our understanding of the communicative abilities of adults with TBI and could lead to a more mechanistic account of the communication difficulties adults with TBI experience in rich communication contexts that require the processing and integration of multiple co-occurring cues. This work has the potential to increase the ecological validity of language assessment and provide insights into the cognitive and neural mechanisms that support multimodal language processing.
语言是多模态的,存在于丰富的视觉语境中。语言也是递增的,在实时中一刻不停地展开,然而,很少有研究考察口语如何在多模态语言处理过程中与手势和视觉语境相互作用。手势是一种丰富的交流线索,与言语紧密相关,通常描绘来自视觉世界的具体指称。我们在改编的视觉世界范式中使用眼动追踪,考察了伴有和不伴有中度至重度创伤性脑损伤(TBI)的参与者如何使用手势来解决暂时的指称模糊。
参与者观看一个带有四个物体和一个视频的屏幕。视频中的说话者说出句子(例如,“女孩将吃非常好的三明治”),同时伴随着有意义的手势(例如,三明治握持手势)或无意义的梳理动作(例如,手臂抓挠)在动词“将吃”上。我们在手势模态中的动作开始与言语中所说的参考物开始之间的关键窗口期间,测量了参与者的注视点目标对象(例如,三明治)、语义竞争者(例如,苹果)和两个不相关的干扰项(例如,钢琴,吉他)。
无论是伴有还是不伴有 TBI 的参与者,当说话者做出手势而不是梳理动作时,更有可能注视目标;然而,与未受伤的参与者相比,TBI 组的效果明显减弱。
我们证明了 TBI 参与者的言语-手势整合能力相对于未受伤的同龄人有所降低。这项研究推进了我们对 TBI 成年人交流能力的理解,并且可以为 TBI 成年人在需要处理和整合多个同时出现的线索的丰富交流环境中遇到的交流困难提供更具机制性的解释。这项工作有可能提高语言评估的生态有效性,并深入了解支持多模态语言处理的认知和神经机制。