Grasping objects in cluttered scenarios is a challenging task in robotics. Performing pre-grasp actions such as pushing and shifting to scatter objects is a way to reduce clutter. Based on deep reinforcement learning, we propose a Fast-Learning Grasping (FLG) framework, that can integrate pre-grasping actions along with grasping to pick up objects from cluttered scenarios with reduced real-world training time. We associate rewards for performing moving actions with the change of environmental clutter and utilize a hybrid triggering method, leading to data-efficient learning and synergy. Then we use the output of an extended fully convolutional network as the value function of each pixel point of the workspace and establish an accurate estimation of the grasp probability for each action. We also introduce a mask function as prior knowledge to enable the agents to focus on the accurate pose adjustment to improve the effectiveness of collecting training data and, hence, to learn efficiently. We carry out pre-training of the FLG over simulated environment, and then the learnt model is transferred to the real world with minimal fine-tuning for further learning during actions. Experimental results demonstrate a 94% grasp success rate and the ability to generalize to novel objects. Compared to state-of-the-art approaches in the literature, the proposed FLG framework can achieve similar or higher grasp success rate with lesser amount of training in the real world. Supplementary video is available at https://youtu.be/e04uDLsxfDg.
翻译:在杂乱的假想中, 抓取对象是机器人的一项艰巨任务。 执行诸如推动和移动以分散物体等预抓动作是减少混乱的一种方法。 根据深层加固学习, 我们提议了一个快速加固( FLG) 框架, 可以将预加增( FLG) 框架整合起来, 同时抓住从杂乱的假想中抓取对象, 减少真实世界培训时间。 我们把执行移动动作的奖励与环境杂乱和使用混合触发方法的变化联系起来, 导致数据高效的学习和协同。 然后, 我们使用扩展的全演化网络的产出作为每个工作空间像素点的值函数, 并准确估计每项行动的把握概率概率。 我们还引入一个掩码功能, 使代理方能够专注于准确的配置调整, 以提高收集培训数据的效率, 从而高效地学习。 我们通过模拟环境对 FLG 进行预先培训, 然后将学习模型转移到现实世界, 微调的高级网络产出, 用于进一步学习行动期间的每个像素点点的值值值值值值, 实验性G- 成功率 以小的FL 成功率, 向新版缩缩缩缩缩缩的校程框架 成功率 成功率 。 以94 成功率 成功率 至新版的F- L 成功率 成功率 成功率 至新版