University of Nebraska-Lincoln, Lincoln, NE, USA.
J Vis. 2021 Jul 6;21(7):9. doi: 10.1167/jov.21.7.9.
Previous attempts to classify task from eye movement data have relied on model architectures designed to emulate theoretically defined cognitive processes and/or data that have been processed into aggregate (e.g., fixations, saccades) or statistical (e.g., fixation density) features. Black box convolutional neural networks (CNNs) are capable of identifying relevant features in raw and minimally processed data and images, but difficulty interpreting these model architectures has contributed to challenges in generalizing lab-trained CNNs to applied contexts. In the current study, a CNN classifier was used to classify task from two eye movement datasets (Exploratory and Confirmatory) in which participants searched, memorized, or rated indoor and outdoor scene images. The Exploratory dataset was used to tune the hyperparameters of the model, and the resulting model architecture was retrained, validated, and tested on the Confirmatory dataset. The data were formatted into timelines (i.e., x-coordinate, y-coordinate, pupil size) and minimally processed images. To further understand the informational value of each component of the eye movement data, the timeline and image datasets were broken down into subsets with one or more components systematically removed. Classification of the timeline data consistently outperformed the image data. The Memorize condition was most often confused with Search and Rate. Pupil size was the least uniquely informative component when compared with the x- and y-coordinates. The general pattern of results for the Exploratory dataset was replicated in the Confirmatory dataset. Overall, the present study provides a practical and reliable black box solution to classifying task from eye movement data.
先前对眼动数据进行任务分类的尝试依赖于旨在模拟理论定义的认知过程和/或已处理成聚合(例如,注视,眼跳)或统计(例如,注视密度)特征的数据的模型架构。黑盒卷积神经网络(CNN)能够识别原始和最小处理数据和图像中的相关特征,但解释这些模型架构的困难导致了将实验室训练的 CNN 推广到应用上下文的挑战。在当前的研究中,使用 CNN 分类器对两个眼动数据集(探索性和验证性)中的任务进行分类,其中参与者搜索、记忆或评估室内和室外场景图像。探索性数据集用于调整模型的超参数,然后对生成的模型架构进行重新训练、验证和在验证性数据集上进行测试。数据被格式化为时间线(即,x 坐标,y 坐标,瞳孔大小)和最小处理的图像。为了进一步了解眼动数据每个组件的信息价值,将时间线和图像数据集分解成一个或多个组件被系统删除的子集。时间线数据的分类始终优于图像数据。与搜索和评分相比,记忆条件最常被混淆。与 x 坐标和 y 坐标相比,瞳孔大小是最不独特的信息组件。探索性数据集的总体结果模式在验证性数据集中得到了复制。总体而言,本研究为从眼动数据中分类任务提供了一种实用且可靠的黑盒解决方案。