Performing closed-loop grasping at close proximity to an object requires a large field of view. However, such images will inevitably bring large amounts of unnecessary background information, especially when the camera is far away from the target object at the initial stage, resulting in performance degradation of the grasping network. To address this problem, we design a novel PEGG-Net, a real-time, pixel-wise, robotic grasp generation network. The proposed lightweight network is inherently able to learn to remove background noise that can reduce grasping accuracy. Our proposed PEGG-Net achieves improved state-of-the-art performance on both Cornell dataset (98.9%) and Jacquard dataset (93.8%). In the real-world tests, PEGG-Net can support closed-loop grasping at up to 50Hz using an image size of 480x480 in dynamic environments. The trained model also generalizes to previously unseen objects with complex geometrical shapes, household objects and workshop tools and achieved an overall grasp success rate of 91.2% in our real-world grasping experiments.
翻译:在接近对象的地方进行闭环捕捉需要大视野。 然而, 这些图像将不可避免地带来大量不必要的背景信息, 尤其是当相机在初始阶段远离目标对象时, 导致抓取网络的性能退化。 为了解决这个问题, 我们设计了一个新型的 PEGG- Net, 一个实时、 像素一样、 机器人抓取生成网络。 拟议的轻量网络在本质上能够学会消除背景噪音, 从而降低抓取精确度。 我们提议的 PEGG- Net 在康奈尔数据集(98.9 % ) 和Jacqurd 数据集(93.8 % ) 上都取得了更好的最新效果。 在现实世界测试中, PEG- Net 可以在动态环境中使用480x480的图像大小支持在高达50赫兹的闭环捕。 经过培训的模型还以复杂的几何形状、 家用物件和车间工具, 并在我们真实世界的捕捉取实验中获得了91. 2%的总体捕捉取率。