Department of Autonomous and Intelligent Systems, Fundación Tekniker, Iñaki Goenaga, 5-20600 Eibar, Spain.
Computer Science and Artificial Intelligence (UPV/EHU), Pº Manuel Lardizabal, 1-20018 Donostia-San Sebastián, Spain.
Sensors (Basel). 2021 Jan 26;21(3):816. doi: 10.3390/s21030816.
Grasping point detection has traditionally been a core robotic and computer vision problem. In recent years, deep learning based methods have been widely used to predict grasping points, and have shown strong generalization capabilities under uncertainty. Particularly, approaches that aim at predicting object affordances without relying on the object identity, have obtained promising results in random bin-picking applications. However, most of them rely on RGB/RGB-D images, and it is not clear up to what extent 3D spatial information is used. Graph Convolutional Networks (GCNs) have been successfully used for object classification and scene segmentation in point clouds, and also to predict grasping points in simple laboratory experimentation. In the present proposal, we adapted the Deep Graph Convolutional Network model with the intuition that learning from -dimensional point clouds would lead to a performance boost to predict object affordances. To the best of our knowledge, this is the first time that GCNs are applied to predict affordances for suction and gripper end effectors in an industrial bin-picking environment. Additionally, we designed a bin-picking oriented data preprocessing pipeline which contributes to ease the learning process and to create a flexible solution for any bin-picking application. To train our models, we created a highly accurate RGB-D/3D dataset which is openly available on demand. Finally, we benchmarked our method against a 2D Fully Convolutional Network based method, improving the top-1 precision score by 1.8% and 1.7% for suction and gripper respectively.
抓取点检测一直是机器人学和计算机视觉的核心问题。近年来,基于深度学习的方法已被广泛用于预测抓取点,并在不确定性下表现出强大的泛化能力。特别是,旨在不依赖物体身份预测物体可及性的方法,在随机分拣应用中取得了有希望的结果。然而,它们大多依赖于 RGB/RGB-D 图像,目前尚不清楚 3D 空间信息的使用程度。图卷积网络(GCN)已成功用于点云中的物体分类和场景分割,以及在简单的实验室实验中预测抓取点。在本提案中,我们基于这样一种直觉来适应深度图卷积网络模型:从三维点云中学习将提高预测物体可及性的性能。据我们所知,这是首次将 GCN 应用于预测工业分拣环境中吸盘和夹持器末端执行器的可及性。此外,我们设计了一种面向分拣的数据集预处理流水线,有助于简化学习过程,并为任何分拣应用提供灵活的解决方案。为了训练我们的模型,我们创建了一个高度准确的 RGB-D/3D 数据集,可根据需要公开获取。最后,我们将我们的方法与基于 2D 全卷积网络的方法进行了基准测试,分别将吸盘和夹持器的 top-1 精度提高了 1.8%和 1.7%。