Gao Hejia, Zhao Junjie, Hu Juqi, Sun Changyin
IEEE Trans Neural Netw Learn Syst. 2025 May;36(5):8215-8226. doi: 10.1109/TNNLS.2024.3419180. Epub 2025 May 2.
In the field of robot grasping detection, due to uncertain factors such as different shapes, distinct colors, diverse materials, and various poses, robot grasping has become very challenging. This article introduces a integrated robotic system designed to address the challenge of grasping numerous unknown objects within a scene from a set of $\alpha $ -channel images. We propose a lightweight and object-independent pixel-level generative adaptive residual depthwise separable convolutional neural network (GARDSCN) with an inference speed of around 28 ms, which can be applied to real-time grasping detection. It can effectively deal with the grasping detection of unknown objects with different shapes and poses in various scenes and overcome the limitations of current robot grasping technology. The proposed network achieves 98.88% grasp detection accuracy on the Cornell dataset and 95.23% on the Jacquard dataset. To further verify the validity, the grasping experiment is conducted on a physical robot Kinova Gen2, and the grasp success rate is 96.67% in the single-object scene and 94.10% in the multiobject cluttered scene.
在机器人抓取检测领域,由于存在不同形状、各异颜色、多样材质和各种姿态等不确定因素,机器人抓取变得极具挑战性。本文介绍了一种集成机器人系统,旨在应对从一组α通道图像中抓取场景内众多未知物体的挑战。我们提出了一种轻量级且与物体无关的像素级生成自适应残差深度可分离卷积神经网络(GARDSCN),其推理速度约为28毫秒,可应用于实时抓取检测。它能够有效处理各种场景中不同形状和姿态的未知物体的抓取检测,并克服当前机器人抓取技术的局限性。所提出的网络在康奈尔数据集上的抓取检测准确率达到98.88%,在提花织物数据集上达到95.23%。为进一步验证其有效性,在物理机器人Kinova Gen2上进行了抓取实验,在单物体场景中的抓取成功率为96.67%,在多物体杂乱场景中为94.10%。