Lavan Nadine, Collins Madeleine Rose Niamh, Miah Jannatul Firdaus Monisha
Department of Biological and Experimental Psychology, School of Biological and Chemical Sciences, Queen Mary University of London, UK.
Br J Psychol. 2022 Feb;113(1):248-263. doi: 10.1111/bjop.12531. Epub 2021 Sep 6.
Identity perception often takes place in multimodal settings, where perceivers have access to both visual (face) and auditory (voice) information. Despite this, identity perception is usually studied in unimodal contexts, where face and voice identity perception are modelled independently from one another. In this study, we asked whether and how much auditory and visual information contribute to audiovisual identity perception from naturally-varying stimuli. In a between-subjects design, participants completed an identity sorting task with either dynamic video-only, audio-only or dynamic audiovisual stimuli. In this task, participants were asked to sort multiple, naturally-varying stimuli from three different people by perceived identity. We found that identity perception was more accurate for video-only and audiovisual stimuli compared with audio-only stimuli. Interestingly, there was no difference in accuracy between video-only and audiovisual stimuli. Auditory information nonetheless played a role alongside visual information as audiovisual identity judgements per stimulus could be predicted from both auditory and visual identity judgements, respectively. While the relationship was stronger for visual information and audiovisual information, auditory information still uniquely explained a significant portion of the variance in audiovisual identity judgements. Our findings thus align with previous theoretical and empirical work that proposes that, compared with faces, voices are an important but relatively less salient and a weaker cue to identity perception. We expand on this work to show that, at least in the context of this study, having access to voices in addition to faces does not result in better identity perception accuracy.
身份感知通常发生在多模态环境中,在这种环境下,感知者可以获取视觉(面部)和听觉(声音)信息。尽管如此,身份感知通常是在单模态环境中进行研究的,在这种环境下,面部和声音身份感知是相互独立建模的。在本研究中,我们探讨了听觉和视觉信息是否以及在多大程度上有助于从自然变化的刺激中进行视听身份感知。在一项被试间设计中,参与者完成了一项身份分类任务,任务中的刺激分别为仅动态视频、仅音频或动态视听刺激。在这个任务中,参与者被要求根据感知到的身份对来自三个不同人的多个自然变化的刺激进行分类。我们发现,与仅音频刺激相比,仅视频和视听刺激的身份感知更准确。有趣的是,仅视频和视听刺激在准确性上没有差异。然而,听觉信息与视觉信息一起发挥了作用,因为每个刺激的视听身份判断可以分别从听觉和视觉身份判断中预测出来。虽然视觉信息和视听信息之间的关系更强,但听觉信息仍然独特地解释了视听身份判断中很大一部分的方差。因此,我们的研究结果与之前的理论和实证研究一致,这些研究表明,与面部相比,声音是身份感知的一个重要但相对不那么突出且较弱的线索。我们扩展了这项研究,以表明至少在本研究的背景下,除了面部之外还能听到声音并不会提高身份感知的准确性。