Salatiello Alessandro, Hovaidi-Ardestani Mohammad, Giese Martin A
Section for Computational Sensomotorics, Department of Cognitive Neurology, Centre for Integrative Neuroscience, Hertie Institute for Clinical Brain Research, University Clinic Tübingen, Tübingen, Germany.
Front Neurorobot. 2021 Jun 9;15:648527. doi: 10.3389/fnbot.2021.648527. eCollection 2021.
The ability to make accurate social inferences makes humans able to navigate and act in their social environment effortlessly. Converging evidence shows that motion is one of the most informative cues in shaping the perception of social interactions. However, the scarcity of parameterized generative models for the generation of highly-controlled stimuli has slowed down both the identification of the most critical motion features and the understanding of the computational mechanisms underlying their extraction and processing from rich visual inputs. In this work, we introduce a novel generative model for the automatic generation of an arbitrarily large number of videos of socially interacting agents for comprehensive studies of social perception. The proposed framework, validated with three psychophysical experiments, allows generating as many as 15 distinct interaction classes. The model builds on classical dynamical system models of biological navigation and is able to generate visual stimuli that are parametrically controlled and representative of a heterogeneous set of social interaction classes. The proposed method represents thus an important tool for experiments aimed at unveiling the computational mechanisms mediating the perception of social interactions. The ability to generate highly-controlled stimuli makes the model valuable not only to conduct behavioral and neuroimaging studies, but also to develop and validate neural models of social inference, and machine vision systems for the automatic recognition of social interactions. In fact, contrasting human and model responses to a heterogeneous set of highly-controlled stimuli can help to identify critical computational steps in the processing of social interaction stimuli.
做出准确社会推断的能力使人类能够在其社会环境中轻松地导航和行动。越来越多的证据表明,运动是塑造社会互动感知的最具信息性的线索之一。然而,用于生成高度可控刺激的参数化生成模型的稀缺,减缓了对最关键运动特征的识别以及对从丰富视觉输入中提取和处理这些特征的计算机制的理解。在这项工作中,我们引入了一种新颖的生成模型,用于自动生成任意数量的社会互动主体视频,以全面研究社会感知。所提出的框架通过三个心理物理学实验得到验证,能够生成多达15种不同的互动类别。该模型基于生物导航的经典动态系统模型构建,能够生成参数化控制的视觉刺激,这些刺激代表了一组异质的社会互动类别。因此,所提出的方法是旨在揭示介导社会互动感知的计算机制的实验的重要工具。生成高度可控刺激的能力使该模型不仅对进行行为和神经成像研究有价值,而且对开发和验证社会推断的神经模型以及用于自动识别社会互动的机器视觉系统也有价值。事实上,对比人类和模型对一组异质的高度可控刺激的反应,有助于识别社会互动刺激处理中的关键计算步骤。