In this paper, we focus on the problem of integrating Energy-based Models (EBM) as guiding priors for motion optimization. EBMs are a set of neural networks that can represent expressive probability density distributions in terms of a Gibbs distribution parameterized by a suitable energy function. Due to their implicit nature, they can easily be integrated as optimization factors or as initial sampling distributions in the motion optimization problem, making them good candidates to integrate data-driven priors in the motion optimization problem. In this work, we present a set of required modeling and algorithmic choices to adapt EBMs into motion optimization. We investigate the benefit of including additional regularizers in the learning of the EBMs to use them with gradient-based optimizers and we present a set of EBM architectures to learn generalizable distributions for manipulation tasks. We present multiple cases in which the EBM could be integrated for motion optimization and evaluate the performance of learned EBMs as guiding priors for both simulated and real robot experiments.
翻译:在本文中,我们侧重于整合以能源为基础的模型(EBM)作为运动优化前导的问题。EBM是一组神经网络,可以代表Gibbs分布参数的显示概率密度分布,以合适的能源功能为参数。由于其隐含性质,它们很容易作为优化因素或运动优化问题的初步抽样分布加以整合,从而使他们在运动优化问题中能够将数据驱动的先导纳入其中。在这项工作中,我们提出了一套必要的模型和算法选择,以将EBM调整为运动优化。我们调查了在EBM的学习中增加正规化器以使用梯度优化器的好处,我们提出了一套EBM结构,以学习用于操作任务的通用分布。我们提出了多种案例,其中EBM可以结合用于运动优化,并评价所学的EBM的性功能,作为模拟和真实机器人实验的前导力。