Inherent morphological characteristics in objects may offer a wide range of plausible grasping orientations that obfuscates the visual learning of robotic grasping. Existing grasp generation approaches are cursed to construct discontinuous grasp maps by aggregating annotations for drastically different orientations per grasping point. Moreover, current methods generate grasp candidates across a single direction in the robot's viewpoint, ignoring its feasibility constraints. In this paper, we propose a novel augmented grasp map representation, suitable for pixel-wise synthesis, that locally disentangles grasping orientations by partitioning the angle space into multiple bins. Furthermore, we introduce the ORientation AtteNtive Grasp synthEsis (ORANGE) framework, that jointly addresses classification into orientation bins and angle-value regression. The bin-wise orientation maps further serve as an attention mechanism for areas with higher graspability, i.e. probability of being an actual grasp point. We report new state-of-the-art 94.71% performance on Jacquard, with a simple U-Net using only depth images, outperforming even multi-modal approaches. Subsequent qualitative results with a real bi-manual robot validate ORANGE's effectiveness in generating grasps for multiple orientations, hence allowing planning grasps that are feasible.
翻译:物体的固有形态特征可能提供一系列令人信服的掌握方向,使机器人掌握的视觉学习模糊不清。 现有的掌握的生成方法被诅咒,通过对每个掌握点截然不同的方向的批注来构建不连贯的掌握的地图。 此外, 目前的方法会从机器人的观点中产生一个单一的方向来抓取候选人, 忽视其可行性限制 。 在本文中, 我们提议了一个新的增强的掌握地图代表, 适合像素合成, 当地通过将角空间分割成多个文件夹来分解捕捉方向。 此外, 我们引入了“ 旋转动动动动引力合成( ORANGE) ” 框架, 该框架共同将分类到方向文件夹和角度值回归。 双向方向的定位图进一步成为机器人更能捕捉的地区的关注机制, 也就是说, 有可能成为实际的掌握点 。 我们在Jacquldard上报告新的状态94.71%的性能, 使用简单的UNet, 仅使用深度图像, 表现甚至多式的方法。 之后的定性结果是, 将真实的双向规划结果, 能够产生真正的双向。