Despite the impressive progress achieved in robotic grasping, robots are not skilled in sophisticated tasks (e.g. search and grasp a specified target in clutter). Such tasks involve not only grasping but the comprehensive perception of the world (e.g. the object relationships). Recently, encouraging results demonstrate that it is possible to understand high-level concepts by learning. However, such algorithms are usually data-intensive, and the lack of data severely limits their performance. In this paper, we present a new dataset named REGRAD for the learning of relationships among objects and grasps. We collect the annotations of object poses, segmentations, grasps, and relationships for the target-driven relational grasping tasks. Our dataset is collected in both forms of 2D images and 3D point clouds. Moreover, since all the data are generated automatically, it is free to import new objects for data generation. We also released a real-world validation dataset to evaluate the sim-to-real performance of models trained on REGRAD. Finally, we conducted a series of experiments, showing that the models trained on REGRAD could generalize well to the realistic scenarios, in terms of both relationship and grasp detection. Our dataset and code could be found at: https://github.com/poisonwine/REGRAD
翻译:尽管在机器人掌握方面取得了令人印象深刻的进展,但机器人在复杂的任务(例如搜索和掌握一个特定的目标)方面技能不足,这些任务不仅涉及掌握,而且涉及对世界的全面认识(例如对象关系)。最近,令人鼓舞的结果表明,通过学习有可能理解高级别概念。然而,这种算法通常是数据密集型的,而且缺乏数据严重限制了其性能。在本文件中,我们提出了一个名为REGRAD的新数据集,用于学习对象和掌握者之间的关系。我们收集了目标驱动关系掌握任务的对象构成、分割、掌握和关系的说明。我们的数据集是以2D图像和3D点云两种形式收集的。此外,由于所有数据都是自动生成的,因此可以自由为数据生成输入新的对象。我们还发布了一个真实世界验证数据集,用于评价在REGRAD上培训的模型的模拟到真实性表现。最后,我们进行了一系列实验,显示在RERAD上培训的模型可以概括到现实的情景,从我们找到的代码/理解到我们的数据。