Imitation learning approaches achieve good generalization within the range of the training data, but tend to generate unpredictable motions when querying outside this range. We present a novel approach to imitation learning with enhanced extrapolation capabilities that exploits the so-called Equation Learner Network (EQLN). Unlike conventional approaches, EQLNs use supervised learning to fit a set of analytical expressions that allows them to extrapolate beyond the range of the training data. We augment the task demonstrations with a set of task-dependent parameters representing spatial properties of each motion and use them to train the EQLN. At run time, the features are used to query the Task-Parameterized Equation Learner Network (TP-EQLN) and generate the corresponding robot trajectory. The set of features encodes kinematic constraints of the task such as desired height or a final point to reach. We validate the results of our approach on manipulation tasks where it is important to preserve the shape of the motion in the extrapolation domain. Our approach is also compared with existing state-of-the-art approaches, in simulation and in real setups. The experimental results show that TP-EQLN can respect the constraints of the trajectory encoded in the feature parameters, even in the extrapolation domain, while preserving the overall shape of the trajectory provided in the demonstrations.
翻译:与常规方法不同, EQLNs 使用监督的学习方法来适应一系列分析表达方式,使其能够超越培训数据的范围进行外推。我们用一组任务参数来扩大任务演示,这些参数代表每个运动的空间特性,并利用它们来培训 EQLN。在运行期间,这些功能被用来查询任务计量的量化学习者网络(TP-EQLN),并生成相应的机器人轨迹。一套特征编码了任务(例如预期的高度或最终达到的点)的有色限制。我们验证了我们处理任务的方法的结果,在操作任务中必须保留外推领域运动的形状。我们的方法也与现有的状态方法相比较,在模拟和真实的域内轨迹演示中,甚至以模拟和真实的轨迹模型形式生成相应的机器人轨迹。