Graduate School of Engineering and Science, University of the Ryukyus, Nishihara 903-0213, Japan.
Department of Engineering, University of the Ryukyus, Nishihara 903-0213, Japan.
Sensors (Basel). 2023 Apr 16;23(8):4028. doi: 10.3390/s23084028.
When an electric wheelchair is operated using gaze motion, eye movements such as checking the environment and observing objects are also incorrectly recognized as input operations. This phenomenon is called the "Midas touch problem", and classifying visual intentions is extremely important. In this paper, we develop a deep learning model that estimates the user's visual intention in real time and an electric wheelchair control system that combines intention estimation and the gaze dwell time method. The proposed model consists of a 1DCNN-LSTM that estimates visual intention from feature vectors of 10 variables, such as eye movement, head movement, and distance to the fixation point. The evaluation experiments classifying four types of visual intentions show that the proposed model has the highest accuracy compared to other models. In addition, the results of the driving experiments of the electric wheelchair implementing the proposed model show that the user's efforts to operate the wheelchair are reduced and that the operability of the wheelchair is improved compared to the traditional method. From these results, we concluded that visual intentions could be more accurately estimated by learning time series patterns from eye and head movement data.
当电动轮椅通过凝视运动操作时,扫视环境和观察物体等眼部运动也可能被错误地识别为输入操作。这种现象被称为“点金手问题”,因此对视觉意图进行分类极为重要。在本文中,我们开发了一种实时估计用户视觉意图的深度学习模型,以及一种结合意图估计和凝视停留时间方法的电动轮椅控制系统。所提出的模型由一个 1DCNN-LSTM 组成,该模型从 10 个变量(如眼球运动、头部运动和与注视点的距离)的特征向量中估计视觉意图。对四种视觉意图进行分类的评估实验表明,与其他模型相比,所提出的模型具有最高的准确性。此外,在实施所提出模型的电动轮椅的驾驶实验结果中,与传统方法相比,用户操作轮椅的努力减少了,轮椅的可操作性提高了。从这些结果中,我们得出结论,通过从眼动和头部运动数据中学习时间序列模式,可以更准确地估计视觉意图。