Rea Francesco, Vignolo Alessia, Sciutti Alessandra, Noceti Nicoletta
Robotics Brain and Cognitive Sciences (RBCS), Istituto Italiano di Tecnologia, Genova, Italy.
CONTACT, Istituto Italiano di Tecnologia, Genova, Italy.
Front Robot AI. 2019 Jul 16;6:58. doi: 10.3389/frobt.2019.00058. eCollection 2019.
In the industry of the future, so as in healthcare and at home, robots will be a familiar presence. Since they will be working closely with human operators not always properly trained for human-machine interaction tasks, robots will need the ability of automatically adapting to changes in the task to be performed or to cope with variations in how the human partner completes the task. The goal of this work is to make a further step toward endowing robot with such capability. To this purpose, we focus on the identification of relevant time instants in an observed action, called , informative on the partner's movement timing, and marking instants where an action starts or ends, or changes to another action. The time instants are temporal locations where the motion can be ideally segmented, providing a set of primitives that can be used to build a temporal signature of the action and finally support the understanding of the dynamics and coordination in time. We validate our approach in two contexts, considering first a situation in which the human partner can perform multiple different activities, and then moving to settings where an action is already recognized and shows a certain degree of periodicity. In the two contexts we address different challenges. In the first one, working in batch on a dataset collecting videos of a variety of cooking activities, we investigate whether the action signature we compute could facilitate the understanding of which type of action is occurring in front of the observer, with tolerance to viewpoint changes. In the second context, we evaluate online on the robot iCub the capability of the action signature in providing hints to establish an actual temporal coordination during the interaction with human participants. In both cases, we show promising results that speak in favor of the potentiality of our approach.
在未来的行业中,就像在医疗保健领域和家庭中一样,机器人将随处可见。由于它们将与并非总是接受过人机交互任务适当培训的人类操作员密切合作,机器人将需要具备自动适应待执行任务变化或应对人类伙伴完成任务方式变化的能力。这项工作的目标是朝着赋予机器人这种能力迈出进一步的步伐。为此,我们专注于识别观察到的动作中的相关时刻,这些时刻被称为对伙伴运动时间有信息价值的时刻,并标记动作开始、结束或转变为另一个动作的时刻。这些时刻是运动可以理想地进行分段的时间位置,提供了一组原语,可用于构建动作的时间特征,最终支持对动态和时间协调的理解。我们在两种情况下验证我们的方法,首先考虑人类伙伴可以执行多种不同活动的情况,然后转向动作已经被识别且呈现一定程度周期性的场景。在这两种情况下,我们应对不同的挑战。在第一种情况下,我们对收集各种烹饪活动视频的数据集进行批量处理,研究我们计算出的动作特征是否有助于理解观察者面前正在发生哪种类型的动作,同时对视角变化具有容忍度。在第二种情况下,我们在机器人iCub上进行在线评估,评估动作特征在与人类参与者交互期间提供线索以建立实际时间协调的能力。在这两种情况下,我们都展示了有前景的结果,这些结果支持我们方法的潜力。