Perceiving and interacting with 3D articulated objects, such as cabinets, doors, and faucets, pose particular challenges for future home-assistant robots performing daily tasks in human environments. Besides parsing the articulated parts and joint parameters, researchers recently advocate learning manipulation affordance over the input shape geometry which is more task-aware and geometrically fine-grained. However, taking only passive observations as inputs, these methods ignore many hidden but important kinematic constraints (e.g., joint location and limits) and dynamic factors (e.g., joint friction and restitution), therefore losing significant accuracy for test cases with such uncertainties. In this paper, we propose a novel framework, named AdaAfford, that learns to perform very few test-time interactions for quickly adapting the affordance priors to more accurate instance-specific posteriors. We conduct large-scale experiments using the PartNet-Mobility dataset and prove that our system performs better than baselines.
翻译:与 3D 表达的物体( 如 柜子、 门和水龙头) 的感知和互动对未来家庭助理机器人在人类环境中执行日常任务构成特殊的挑战。 除了解析表达的部件和联合参数外,研究人员最近还倡导对输入形状几何进行学习操纵,这种几何更具有任务意识和几何精细的特性。然而,这些方法只以被动观察作为输入,忽略了许多隐藏但重要的运动障碍(如联合位置和限制)和动态因素(如联合摩擦和复原),从而丧失了具有这种不确定性的测试案例的显著准确性。 在本文中,我们提出了一个名为AdaAffford的新框架,它学会进行很少的测试时间互动,以便快速调整前的发价,使其适应更精确的外表。我们使用 PartNet- 移动数据集进行大规模实验, 并证明我们的系统比基线要好。