Küntzler Theresa, Höfling T Tim A, Alpers Georg W
Department of Politics and Public Administration, Center for Image Analysis in the Social Sciences, Graduate School of Decision Science, University of Konstanz, Konstanz, Germany.
Department of Psychology, School of Social Sciences, University of Mannheim, Mannheim, Germany.
Front Psychol. 2021 May 5;12:627561. doi: 10.3389/fpsyg.2021.627561. eCollection 2021.
Emotional facial expressions can inform researchers about an individual's emotional state. Recent technological advances open up new avenues to automatic Facial Expression Recognition (FER). Based on machine learning, such technology can tremendously increase the amount of processed data. FER is now easily accessible and has been validated for the classification of standardized prototypical facial expressions. However, applicability to more naturalistic facial expressions still remains uncertain. Hence, we test and compare performance of three different FER systems (Azure Face API, Microsoft; Face++, Megvii Technology; FaceReader, Noldus Information Technology) with human emotion recognition (A) for standardized posed facial expressions (from prototypical inventories) and (B) for non-standardized acted facial expressions (extracted from emotional movie scenes). For the standardized images, all three systems classify basic emotions accurately (FaceReader is most accurate) and they are mostly on par with human raters. For the non-standardized stimuli, performance drops remarkably for all three systems, but Azure still performs similarly to humans. In addition, all systems and humans alike tend to misclassify some of the non-standardized emotional facial expressions as neutral. In sum, emotion recognition by automated facial expression recognition can be an attractive alternative to human emotion recognition for standardized and non-standardized emotional facial expressions. However, we also found limitations in accuracy for specific facial expressions; clearly there is need for thorough empirical evaluation to guide future developments in computer vision of emotional facial expressions.
情绪性面部表情能够让研究人员了解个体的情绪状态。近期的技术进步为自动面部表情识别(FER)开辟了新途径。基于机器学习的此类技术能够极大地增加处理数据的量。如今FER很容易获取,并且已被验证可用于标准化原型面部表情的分类。然而,其对更自然面部表情的适用性仍不确定。因此,我们测试并比较了三种不同FER系统(微软的Azure Face API、旷视科技的Face++、诺达思信息技术公司的FaceReader)与人类情绪识别在以下两种情况下的表现:(A)针对标准化摆拍面部表情(来自原型库),以及(B)针对非标准化表演面部表情(从情感电影场景中提取)。对于标准化图像,所有这三个系统都能准确地对面部基本情绪进行分类(FaceReader最为准确),并且它们大多与人类评分者的表现相当。对于非标准化刺激,这三个系统的表现都显著下降,但Azure的表现仍与人类相似。此外,所有系统以及人类都倾向于将一些非标准化情绪性面部表情误分类为中性表情。总之,对于标准化和非标准化情绪性面部表情而言,通过自动面部表情识别进行情绪识别可能是人类情绪识别的一个有吸引力的替代方法。然而,我们也发现了特定面部表情在准确性方面存在的局限性;显然需要进行全面的实证评估来指导情绪性面部表情计算机视觉的未来发展。