Department of Mathematics and Informatics, Universitat de Barcelona, 08007 Barcelona, Spain.
Computer Vision Center, Campus UAB, 08193 Bellaterra, Spain.
Sensors (Basel). 2021 Jul 13;21(14):4769. doi: 10.3390/s21144769.
This paper summarizes the OpenEDS 2020 Challenge dataset, the proposed baselines, and results obtained by the top three winners of each competition: (1) Gaze prediction Challenge, with the goal of predicting the gaze vector 1 to 5 frames into the future based on a sequence of previous eye images, and (2) Sparse Temporal Semantic Segmentation Challenge, with the goal of using temporal information to propagate semantic eye labels to contiguous eye image frames. Both competitions were based on the OpenEDS2020 dataset, a novel dataset of eye-image sequences captured at a frame rate of 100 Hz under controlled illumination, using a virtual-reality head-mounted display with two synchronized eye-facing cameras. The dataset, which we make publicly available for the research community, consists of 87 subjects performing several gaze-elicited tasks, and is divided into 2 subsets, one for each competition task. The proposed baselines, based on deep learning approaches, obtained an average angular error of 5.37 degrees for gaze prediction, and a mean intersection over union score (mIoU) of 84.1% for semantic segmentation. The winning solutions were able to outperform the baselines, obtaining up to 3.17 degrees for the former task and 95.2% mIoU for the latter.
本文总结了 OpenEDS 2020 挑战赛数据集、提出的基线以及每个竞赛的前三名获胜者的结果:(1)注视预测挑战赛,目标是根据一系列先前的眼图像预测未来 1 到 5 帧的注视向量;(2)稀疏时间语义分割挑战赛,目标是利用时间信息将语义眼标签传播到连续的眼图像帧。这两个竞赛都是基于 OpenEDS2020 数据集进行的,该数据集是在受控照明下以 100Hz 的帧率捕获的眼图像序列的新型数据集,使用具有两个同步眼对置相机的虚拟现实头戴式显示器。我们为研究社区公开提供了该数据集,其中包含 87 位受试者执行的多项注视诱发任务,并分为两个子集,每个子集用于一个竞赛任务。基于深度学习方法的提出的基线在注视预测方面获得了 5.37 度的平均角度误差,在语义分割方面获得了 84.1%的平均交并比(mIoU)得分。获胜的解决方案能够超越基线,在前一项任务中最高可达 3.17 度,在后一项任务中最高可达 95.2%的 mIoU。