Imitation Learning (IL) holds great potential for learning repetitive manipulation tasks, such as those in industrial assembly. However, its effectiveness is often limited by insufficient trajectory precision due to compounding errors. In this paper, we introduce Grasped Object Manifold Projection (GrOMP), an interactive method that mitigates these errors by constraining a non-rigidly grasped object to a lower-dimensional manifold. GrOMP assumes a precise task in which a manipulator holds an object that may shift within the grasp in an observable manner and must be mated with a grounded part. Crucially, all GrOMP enhancements are learned from the same expert dataset used to train the base IL policy, and are adjusted with an n-arm bandit-based interactive component. We propose a theoretical basis for GrOMP's improvement upon the well-known compounding error bound in IL literature. We demonstrate the framework on four precise assembly tasks using tactile feedback, and note that the approach remains modality-agnostic. Data and videos are available at williamvdb.github.io/GrOMPsite.
翻译:模仿学习(IL)在工业装配等重复性操作任务的学习中具有巨大潜力。然而,由于误差累积导致的轨迹精度不足,其效果常受限制。本文提出抓取物体流形投影(GrOMP),这是一种通过将非刚性抓取的物体约束到低维流形来缓解此类误差的交互式方法。GrOMP适用于一类精密任务:机械臂抓持的物体在可观测范围内可能发生抓持偏移,且需与接地部件进行配合。关键的是,所有GrOMP增强模块均从训练基础IL策略所用的同一专家数据集中学习,并通过基于n臂赌博机的交互组件进行调节。我们为GrOMP对IL文献中经典误差累积界限的改进提供了理论依据。该框架在四个利用触觉反馈的精密装配任务中得到验证,且方法本身保持模态无关性。数据与视频详见williamvdb.github.io/GrOMPsite。