Computer Science Department, Open University, Raanana, Israel.
Information Systems Department, University of Haifa, Haifa, Israel.
Sci Rep. 2022 Dec 30;12(1):22611. doi: 10.1038/s41598-022-27079-w.
In animal research, automation of affective states recognition has so far mainly addressed pain in a few species. Emotional states remain uncharted territories, especially in dogs, due to the complexity of their facial morphology and expressions. This study contributes to fill this gap in two aspects. First, it is the first to address dog emotional states using a dataset obtained in a controlled experimental setting, including videos from (n = 29) Labrador Retrievers assumed to be in two experimentally induced emotional states: negative (frustration) and positive (anticipation). The dogs' facial expressions were measured using the Dogs Facial Action Coding System (DogFACS). Two different approaches are compared in relation to our aim: (1) a DogFACS-based approach with a two-step pipeline consisting of (i) a DogFACS variable detector and (ii) a positive/negative state Decision Tree classifier; (2) An approach using deep learning techniques with no intermediate representation. The approaches reach accuracy of above 71% and 89%, respectively, with the deep learning approach performing better. Secondly, this study is also the first to study explainability of AI models in the context of emotion in animals. The DogFACS-based approach provides decision trees, that is a mathematical representation which reflects previous findings by human experts in relation to certain facial expressions (DogFACS variables) being correlates of specific emotional states. The deep learning approach offers a different, visual form of explainability in the form of heatmaps reflecting regions of focus of the network's attention, which in some cases show focus clearly related to the nature of particular DogFACS variables. These heatmaps may hold the key to novel insights on the sensitivity of the network to nuanced pixel patterns reflecting information invisible to the human eye.
在动物研究中,情感状态的自动识别迄今为止主要针对少数几种动物的疼痛。由于其面部形态和表情的复杂性,情感状态仍然是未知领域,尤其是在狗身上。本研究从两个方面有助于填补这一空白。首先,它是第一个使用在受控实验环境中获得的数据集来解决狗的情感状态的问题,该数据集包括(n=29) Labrador Retrievers 的视频,这些狗被认为处于两种实验诱导的情绪状态:消极(挫折)和积极(期待)。狗的面部表情使用狗面部动作编码系统(DogFACS)进行测量。两种不同的方法与我们的目标进行了比较:(1)一种基于 DogFACS 的方法,该方法具有两步流水线,包括(i)DogFACS 变量检测器和(ii)正/负状态决策树分类器;(2)一种使用深度学习技术的方法,没有中间表示。这两种方法的准确率分别达到了 71%以上和 89%以上,深度学习方法的表现更好。其次,本研究也是第一个研究动物情感背景下人工智能模型的可解释性的研究。基于 DogFACS 的方法提供了决策树,即一种数学表示形式,反映了人类专家在某些面部表情(DogFACS 变量)与特定情绪状态相关时的先前发现。深度学习方法提供了一种不同的、可视化的可解释性形式,以热力图的形式反映了网络关注的焦点区域,在某些情况下,这些区域的焦点与特定 DogFACS 变量的性质明显相关。这些热力图可能为网络对反映人眼不可见信息的细微像素模式的敏感性提供了新的见解。