Han Jane, Chauhan Vassiki, Philip Rebecca, Taylor Morgan K, Jung Heejung, Halchenko Yaroslav O, Gobbini M Ida, Haxby James V, Nastase Samuel A
bioRxiv. 2024 Nov 26:2024.11.26.624178. doi: 10.1101/2024.11.26.624178.
We effortlessly extract behaviorally relevant information from dynamic visual input in order to understand the actions of others. In the current study, we develop and test a number of models to better understand the neural representational geometries supporting action understanding. Using fMRI, we measured brain activity as participants viewed a diverse set of 90 different video clips depicting social and nonsocial actions in real-world contexts. We developed five behavioral models using arrangement tasks: two models reflecting behavioral judgments of the purpose (transitivity) and the social content (sociality) of the actions depicted in the video stimuli; and three models reflecting behavioral judgments of the visual content (people, objects, and scene) depicted in still frames of the stimuli. We evaluated how well these models predict neural representational geometry and tested them against semantic models based on verb and nonverb embeddings and visual models based on gaze and motion energy. Our results revealed that behavioral judgments of similarity better reflect neural representational geometry than semantic or visual models throughout much of cortex. The sociality and transitivity models in particular captured a large portion of unique variance throughout the action observation network, extending into regions not typically associated with action perception, like ventral temporal cortex. Overall, our findings expand the action observation network and indicate that the social content and purpose of observed actions are predominant in cortical representation.
我们能轻松地从动态视觉输入中提取与行为相关的信息,以便理解他人的行为。在当前的研究中,我们开发并测试了一些模型,以更好地理解支持动作理解的神经表征几何结构。利用功能磁共振成像(fMRI),我们测量了参与者观看90个不同视频片段时的大脑活动,这些片段描绘了现实世界背景下的社会和非社会行为。我们使用排列任务开发了五个行为模型:两个模型反映对视频刺激中所描绘动作的目的(传递性)和社会内容(社会性)的行为判断;三个模型反映对刺激静止帧中所描绘视觉内容(人物、物体和场景)的行为判断。我们评估了这些模型对神经表征几何结构的预测能力,并将它们与基于动词和非动词嵌入的语义模型以及基于注视和运动能量的视觉模型进行了比较。我们的结果表明,在整个大脑皮层的大部分区域,相似性的行为判断比语义或视觉模型能更好地反映神经表征几何结构。特别是社会性和传递性模型在整个动作观察网络中捕捉到了很大一部分独特的方差,延伸到了通常与动作感知无关的区域,如腹侧颞叶皮层。总体而言,我们的研究结果扩展了动作观察网络,并表明观察到的动作的社会内容和目的在皮层表征中占主导地位。