Department of Experimental Psychology, University College London, 26 Bedford Way, London, WC1H 0AP, UK.
University of Bremen, Bremen, Germany.
Behav Res Methods. 2021 Apr;53(2):686-701. doi: 10.3758/s13428-020-01443-y.
With a shift in interest toward dynamic expressions, numerous corpora of dynamic facial stimuli have been developed over the past two decades. The present research aimed to test existing sets of dynamic facial expressions (published between 2000 and 2015) in a cross-corpus validation effort. For this, 14 dynamic databases were selected that featured facial expressions of the basic six emotions (anger, disgust, fear, happiness, sadness, surprise) in posed or spontaneous form. In Study 1, a subset of stimuli from each database (N = 162) were presented to human observers and machine analysis, yielding considerable variance in emotion recognition performance across the databases. Classification accuracy further varied with perceived intensity and naturalness of the displays, with posed expressions being judged more accurately and as intense, but less natural compared to spontaneous ones. Study 2 aimed for a full validation of the 14 databases by subjecting the entire stimulus set (N = 3812) to machine analysis. A FACS-based Action Unit (AU) analysis revealed that facial AU configurations were more prototypical in posed than spontaneous expressions. The prototypicality of an expression in turn predicted emotion classification accuracy, with higher performance observed for more prototypical facial behavior. Furthermore, technical features of each database (i.e., duration, face box size, head rotation, and motion) had a significant impact on recognition accuracy. Together, the findings suggest that existing databases vary in their ability to signal specific emotions, thereby facing a trade-off between realism and ecological validity on the one end, and expression uniformity and comparability on the other.
随着人们对动态表情的兴趣发生转变,在过去的二十年中,已经开发出了许多动态面部刺激物的语料库。本研究旨在通过跨语料库验证来测试现有的动态面部表情集(发表于 2000 年至 2015 年之间)。为此,选择了 14 个动态数据库,这些数据库以呈现或自然形式展示了基本的六种情绪(愤怒、厌恶、恐惧、快乐、悲伤、惊讶)的面部表情。在研究 1 中,从每个数据库中选择了刺激子集(N=162),呈现给人类观察者和机器分析,从而在数据库之间产生了相当大的情感识别性能差异。分类准确性还进一步因显示的感知强度和自然度而异,与自然的表情相比,呈现的表情被判断更准确、更强烈,但不太自然。研究 2 通过对整个刺激集(N=3812)进行机器分析,旨在对 14 个数据库进行全面验证。基于 FACS 的动作单元(AU)分析表明,在呈现的表情中,面部 AU 配置比自然的表情更具原型性。表情的原型性反过来又预测了情绪分类的准确性,具有更高的原型性的面部行为观察到了更高的性能。此外,每个数据库的技术特征(即持续时间、面部框大小、头部旋转和运动)对识别准确性有重大影响。总之,这些发现表明,现有的数据库在表达特定情感的能力方面存在差异,因此,一方面存在真实性和生态有效性之间的权衡,另一方面存在表达一致性和可比性之间的权衡。