Bhat Ajaz A, Mohan Vishwanathan
1School of Psychology, University of East Anglia, Norwich, UK.
2CSEE Department, University of Essex, Colchester, UK.
Cognit Comput. 2018;10(4):558-576. doi: 10.1007/s12559-018-9553-1. Epub 2018 Apr 14.
From social dining in households to product assembly in manufacturing lines, goal-directed reasoning and cooperation with other agents in shared workspaces is a ubiquitous aspect of our day-to-day activities. Critical for such behaviours is the ability to spontaneously anticipate what is doable by oneself as well as the interacting partner based on the evolving environmental context and thereby exploit such information to engage in goal-oriented action sequences. In the setting of an industrial task where two robots are jointly assembling objects in a shared workspace, we describe a bioinspired neural architecture for goal-directed action planning based on coupled interactions between multiple internal models, primarily of the robot's body and its peripersonal space. The internal models (of each robot's body and peripersonal space) are learnt through a process of sensorimotor exploration and then employed in a range of anticipations related to the feasibility and consequence of potential actions of two industrial robots in the context of a joint goal. The ensuing behaviours are demonstrated in a real-world industrial scenario where two robots are assembling industrial fuse-boxes from multiple constituent objects (fuses, fuse-stands) scattered randomly in their workspace. In a spatially unstructured and temporally evolving assembly scenario, the robots employ reward-based dynamics to plan and anticipate which objects to act on at what time instances so as to successfully complete as many assemblies as possible. The existing spatial setting fundamentally necessitates planning collision-free trajectories and avoiding potential collisions between the robots. Furthermore, an interesting scenario where the assembly goal is not realizable by either of the robots individually but only realizable if they meaningfully cooperate is used to demonstrate the interplay between perception, simulation of multiple internal models and the resulting complementary goal-directed actions of both robots. Finally, the proposed neural framework is benchmarked against a typically engineered solution to evaluate its performance in the assembly task. The framework provides a computational outlook to the emerging results from neurosciences related to the learning and use of body schema and peripersonal space for embodied simulation of action and prediction. While experiments reported here engage the architecture in a complex planning task specifically, the internal model based framework is domain-agnostic facilitating portability to several other tasks and platforms.
从家庭中的社交聚餐到生产线中的产品组装,在共享工作空间中进行目标导向的推理以及与其他主体合作是我们日常活动中普遍存在的一个方面。对于此类行为至关重要的是,能够根据不断变化的环境背景自发地预测自己以及互动伙伴可以做什么,从而利用这些信息参与目标导向的行动序列。在一个工业任务场景中,两个机器人在共享工作空间中共同组装物体,我们描述了一种受生物启发的神经架构,用于基于多个内部模型之间的耦合交互进行目标导向的行动规划,这些内部模型主要涉及机器人的身体及其周边空间。(每个机器人的身体和周边空间的)内部模型通过感觉运动探索过程来学习,然后应用于一系列与两个工业机器人在共同目标背景下潜在行动的可行性和后果相关的预测中。在一个真实的工业场景中展示了由此产生的行为,在该场景中,两个机器人从随机散落在其工作空间中的多个组成物体(保险丝、保险丝座)组装工业保险丝盒。在一个空间上无结构且时间上不断演变的组装场景中,机器人采用基于奖励的动力学来规划和预测在什么时刻对哪些物体采取行动,以便尽可能成功地完成尽可能多的组装任务。现有的空间设置从根本上要求规划无碰撞轨迹并避免机器人之间的潜在碰撞。此外,还使用了一个有趣的场景,即单个机器人都无法实现组装目标,只有当它们进行有意义的合作时才能实现,以此来展示感知、多个内部模型的模拟以及两个机器人由此产生的互补性目标导向行动之间的相互作用。最后,将所提出的神经框架与一种典型的工程解决方案进行基准测试,以评估其在组装任务中的性能。该框架为神经科学中与身体图式和周边空间的学习及使用相关的新兴结果提供了一个计算视角,用于对行动和预测进行具身模拟。虽然这里报告的实验专门让该架构参与一个复杂的规划任务,但基于内部模型的框架与领域无关,便于移植到其他几个任务和平台。