A kitchen assistant needs to operate human-scale objects, such as cabinets and ovens, in unmapped environments with dynamic obstacles. Autonomous interactions in such real-world environments require integrating dexterous manipulation and fluid mobility. While mobile manipulators in different form-factors provide an extended workspace, their real-world adoption has been limited. This limitation is in part due to two main reasons: 1) inability to interact with unknown human-scale objects such as cabinets and ovens, and 2) inefficient coordination between the arm and the mobile base. Executing a high-level task for general objects requires a perceptual understanding of the object as well as adaptive whole-body control among dynamic obstacles. In this paper, we propose a two-stage architecture for autonomous interaction with large articulated objects in unknown environments. The first stage uses a learned model to estimate the articulated model of a target object from an RGB-D input and predicts an action-conditional sequence of states for interaction. The second stage comprises of a whole-body motion controller to manipulate the object along the generated kinematic plan. We show that our proposed pipeline can handle complicated static and dynamic kitchen settings. Moreover, we demonstrate that the proposed approach achieves better performance than commonly used control methods in mobile manipulation. For additional material, please check: https://www.pair.toronto.edu/articulated-mm/ .
翻译:厨房助理需要在一个没有绘图的环境里操作人类规模的物体,例如有动态障碍的柜子和烤箱; 在这种现实环境中的自主互动需要集成式的操纵和流体机动性; 不同形式因素的移动操纵器提供扩展的工作空间,但其真实世界的采用有限。 这一限制部分是由于两个主要原因:(1) 无法与未知的人类规模的物体(如柜子和烤箱)互动, 以及(2) 手臂和移动基地之间协调效率低下。 执行一般物体的高层次任务需要对物体的感知性理解以及动态障碍之间的适应性整体机能控制。 在本文中,我们提出一个与不明环境中的大型表达物体进行自主互动的两阶段结构。 第一阶段使用一个学习模型来估计 RGB-D 投入中的目标物体的阐述模型,并预测国家互动的行动条件序列。 第二阶段包括一个全体运动控制器, 以按照生成的运动计划操纵物体。 我们表明,我们提议的管道可以处理复杂的静态/动态在未知的环境中进行移动操纵。 此外,我们用一个学习模型来评估。 用于常规/动态的操作。