University of Florence, Department of Industrial Engineering, Florence, 56139, Italy.
Scuola Superiore Sant'Anna, The BioRobotics Institute, Pontedera, PI, 56025, Italy.
Sci Data. 2022 May 18;9(1):218. doi: 10.1038/s41597-022-01324-3.
This paper makes the VISTA database, composed of inertial and visual data, publicly available for gesture and activity recognition. The inertial data were acquired with the SensHand, which can capture the movement of wrist, thumb, index and middle fingers, while the RGB-D visual data were acquired simultaneously from two different points of view, front and side. The VISTA database was acquired in two experimental phases: in the former, the participants have been asked to perform 10 different actions; in the latter, they had to execute five scenes of daily living, which corresponded to a combination of the actions of the selected actions. In both phase, Pepper interacted with participants. The two camera point of views mimic the different point of view of pepper. Overall, the dataset includes 7682 action instances for the training phase and 3361 action instances for the testing phase. It can be seen as a framework for future studies on artificial intelligence techniques for activity recognition, including inertial-only data, visual-only data, or a sensor fusion approach.
本文构建了一个包含惯性和视觉数据的 VISTA 数据库,用于手势和活动识别。惯性数据由 SensHand 采集,它可以捕捉手腕、拇指、食指和中指的运动,而 RGB-D 视觉数据则从两个不同的视角(正面和侧面)同时采集。VISTA 数据库是在两个实验阶段采集的:在前者中,参与者被要求执行 10 种不同的动作;在后者中,他们必须执行五个日常生活场景,这些场景对应于所选动作的组合。在两个阶段中,Pepper 都与参与者进行了交互。两个摄像机视角模拟了 Pepper 的不同视角。总的来说,该数据集包括训练阶段的 7682 个动作实例和测试阶段的 3361 个动作实例。它可以被视为未来基于人工智能技术的活动识别研究的框架,包括仅使用惯性数据、仅使用视觉数据或传感器融合方法。