Two main approaches have been developed for learning first-order planning (action) models from unstructured data: combinatorial approaches that yield crisp action schemas from the structure of the state space, and deep learning approaches that produce action schemas from states represented by images. A benefit of the former approach is that the learned action schemas are similar to those that can be written by hand; a benefit of the latter is that the learned representations (predicates) are grounded on the images, and as a result, new instances can be given in terms of images. In this work, we develop a new formulation for learning crisp first-order planning models that are grounded on parsed images, a step to combine the benefits of the two approaches. Parsed images are assumed to be given in a simple O2D language (objects in 2D) that involves a small number of unary and binary predicates like "left", "above", "shape", etc. After learning, new planning instances can be given in terms of pairs of parsed images, one for the initial situation and the other for the goal. Learning and planning experiments are reported for several domains including Blocks, Sokoban, IPC Grid, and Hanoi.
翻译:为了从非结构化数据中学习一阶规划(行动)模型,已经开发了两种主要方法:组合式方法,从国家空间结构中产生精确的一阶规划模型,以及深层次学习方法,从图像所代表的国家产生行动模型。前一种方法的一个好处是,所学的一阶规划方法与手写方法相似;后者的一个好处是,所学的一阶规划方法(预测)以图像为基础,因此,可以提供图像方面的新的实例。在这项工作中,我们为学习以剖析图像为基础的一阶规划模型制定了新的配方,这是将两种方法的惠益结合起来的一个步骤。包建图像假定以简单的O2D语言(2D中的对象)提供,其中涉及少量的“左”、“下方”、“阴影”、“形状”等的单线和二元前导,等等。在学习后,可以提供一对相图像方面的新的规划实例,一个是初步图像,一个是初步情况,另一个是目标的一阶,这是将两种方法的好处结合起来的一个步骤。据假设用一种简单的OD语言(2D)提供,包括数个磁区和磁区。