Division of Electronics and Electrical Engineering, Dongguk University, 30 Pildong-ro 1-gil, Jung-gu, Seoul 100-715, Korea.
Sensors (Basel). 2019 Jan 7;19(1):197. doi: 10.3390/s19010197.
Studies are being actively conducted on camera-based driver gaze tracking in a vehicle environment for vehicle interfaces and analyzing forward attention for judging driver inattention. In existing studies on the single-camera-based method, there are frequent situations in which the eye information necessary for gaze tracking cannot be observed well in the camera input image owing to the turning of the driver's head during driving. To solve this problem, existing studies have used multiple-camera-based methods to obtain images to track the driver's gaze. However, this method has the drawback of an excessive computation process and processing time, as it involves detecting the eyes and extracting the features of all images obtained from multiple cameras. This makes it difficult to implement it in an actual vehicle environment. To solve these limitations of existing studies, this study proposes a method that uses a shallow convolutional neural network (CNN) for the images of the driver's face acquired from two cameras to adaptively select camera images more suitable for detecting eye position; faster R-CNN is applied to the selected driver images, and after the driver's eyes are detected, the eye positions of the camera image of the other side are mapped through a geometric transformation matrix. Experiments were conducted using the self-built Dongguk Dual Camera-based Driver Database (DDCD-DB1) including the images of 26 participants acquired from inside a vehicle and the Columbia Gaze Data Set (CAVE-DB) open database. The results confirmed that the performance of the proposed method is superior to those of the existing methods.
在车辆环境中,基于摄像头的驾驶员注视跟踪技术正在被积极研究,用于车辆界面,并分析驾驶员的前向注意力以判断其是否分心。在基于单摄像头的现有研究中,由于驾驶员在驾驶过程中转头,摄像头输入图像中经常无法很好地观察到进行注视跟踪所需的眼部信息。为了解决这个问题,现有研究使用了基于多摄像头的方法来获取图像以跟踪驾驶员的注视。然而,这种方法存在计算过程和处理时间过长的缺点,因为它涉及到检测眼睛并提取来自多个摄像头获取的所有图像的特征。这使得在实际车辆环境中实现它变得困难。为了解决现有研究的这些局限性,本研究提出了一种使用两个摄像头获取的驾驶员面部图像的浅层卷积神经网络 (CNN) 自适应地选择更适合检测眼睛位置的摄像头图像的方法;将更快的 R-CNN 应用于所选驾驶员图像,在检测到驾驶员的眼睛后,通过几何变换矩阵将另一侧的摄像头图像的眼睛位置映射出来。实验使用了自建的包括 26 名参与者在车内采集的图像的 Dongguk 双摄像头驾驶员数据库 (DDCD-DB1) 和公开的 Columbia Gaze 数据集 (CAVE-DB) 进行。结果证实,所提出方法的性能优于现有方法。