Amsterdam UMC, University of Amsterdam, Department of Medical Psychology, Amsterdam Public Health, Location AMC, Meibergdreef 9, 1100 DD, Amsterdam, The Netherlands.
PSI-EAVISE, Electrical Engineering Technology (ESAT), KU Leuven, De Nayer Campus, Sint-Katelijne-Waver, Belgium.
Behav Res Methods. 2021 Oct;53(5):2037-2048. doi: 10.3758/s13428-021-01544-2. Epub 2021 Mar 19.
The assessment of gaze behaviour is essential for understanding the psychology of communication. Mobile eye-tracking glasses are useful to measure gaze behaviour during dynamic interactions. Eye-tracking data can be analysed by using manually annotated areas-of-interest. Computer vision algorithms may alternatively be used to reduce the amount of manual effort, but also the subjectivity and complexity of these analyses. Using additional re-identification (Re-ID) algorithms, different participants in the interaction can be distinguished. The aim of this study was to compare the results of manual annotation of mobile eye-tracking data with the results of a computer vision algorithm. We selected the first minute of seven randomly selected eye-tracking videos of consultations between physicians and patients in a Dutch Internal Medicine out-patient clinic. Three human annotators and a computer vision algorithm annotated mobile eye-tracking data, after which interrater reliability was assessed between the areas-of-interest annotated by the annotators and the computer vision algorithm. Additionally, we explored interrater reliability when using lengthy videos and different area-of-interest shapes. In total, we analysed more than 65 min of eye-tracking videos manually and with the algorithm. Overall, the absolute normalized difference between the manual and the algorithm annotations of face-gaze was less than 2%. Our results show high interrater agreements between human annotators and the algorithm with Cohen's kappa ranging from 0.85 to 0.98. We conclude that computer vision algorithms produce comparable results to those of human annotators. Analyses by the algorithm are not subject to annotator fatigue or subjectivity and can therefore advance eye-tracking analyses.
注视行为评估对于理解交流心理学至关重要。移动眼动追踪眼镜可用于测量动态交互过程中的注视行为。可以使用手动标注的兴趣区域来分析眼动追踪数据。或者可以使用计算机视觉算法来减少手动工作量,同时降低分析的主观性和复杂性。通过使用附加的重新识别(Re-ID)算法,可以区分交互中的不同参与者。本研究的目的是比较手动标注移动眼动追踪数据的结果与计算机视觉算法的结果。我们选择了荷兰内科门诊中医生和患者之间的七次随机眼动追踪视频中的前一分钟。三位人类注释者和一个计算机视觉算法对移动眼动追踪数据进行了注释,然后评估了注释者和计算机视觉算法注释的兴趣区域之间的评分者间可靠性。此外,我们还探讨了使用冗长的视频和不同的兴趣区域形状时的评分者间可靠性。总共手动和通过算法分析了超过 65 分钟的眼动追踪视频。总体而言,手动和算法标注的面部注视之间的绝对归一化差异小于 2%。我们的结果表明,人类注释者和算法之间的评分者间一致性很高,Cohen's kappa 范围从 0.85 到 0.98。我们得出结论,计算机视觉算法可产生与人类注释者相当的结果。算法分析不受注释者疲劳或主观性的影响,因此可以促进眼动追踪分析。