Annu Int Conf IEEE Eng Med Biol Soc. 2021 Nov;2021:4796-4800. doi: 10.1109/EMBC46164.2021.9630471.
Gaze-based interfaces are especially useful for people with disabilities involving the upper limbs or hands. Typically, users select from a number of options (e.g. letters or commands) displayed on a screen by gazing at the desired option. However, in some applications, e.g. gaze-based driving, it may be dangerous to direct gaze away from the environment towards a separate display. In addition, a purely gaze based interface can present a high cognitive load to users, as gaze is not normally used for selection and/or control, but rather for other purposes, such as information gathering. To address these issues, this paper presents a cost-effective multi-modal system for gaze based driving which combines appearance-based gaze estimates derived from webcam images with push button inputs that trigger command execution. This system uses an intuitive "direct interface", where users determine the direction of motion by gazing in the corresponding direction in the environment. We have implemented the system for both wheelchair control and robotic teleoperation. The use of our system should provide substantial benefits for patients with severe motor disabilities, such as ALS, by providing them with a more natural and affordable method of wheelchair control. We compare the performance of our system to the more conventional and common "indirect" system where gaze is used to select commands from a separate display, showing that our system enables faster and more efficient navigation.
基于注视的界面对于涉及上肢或手部的残疾人士特别有用。通常,用户通过注视所需的选项来从屏幕上显示的多个选项(例如字母或命令)中进行选择。但是,在某些应用程序(例如基于注视的驾驶)中,将目光从环境转移到单独的显示器可能会很危险。此外,纯粹基于注视的界面可能会给用户带来很高的认知负担,因为注视通常不用于选择和/或控制,而是用于其他目的,例如信息收集。为了解决这些问题,本文提出了一种具有成本效益的基于注视的驾驶多模态系统,该系统将基于网络摄像头图像的基于外观的注视估计与触发命令执行的按钮输入相结合。该系统使用直观的“直接界面”,用户通过在环境中相应的方向注视来确定运动方向。我们已经为轮椅控制和机器人遥操作实现了该系统。我们的系统应该为患有严重运动障碍(例如 ALS)的患者提供实质性的好处,因为它为他们提供了一种更自然、更经济实惠的轮椅控制方法。我们将我们的系统与更传统和常见的“间接”系统进行了性能比较,其中注视用于从单独的显示器中选择命令,结果表明我们的系统能够实现更快、更高效的导航。