Current domain-independent, classical planners require symbolic models of the problem domain and instance as input, resulting in a knowledge acquisition bottleneck. Meanwhile, although deep learning has achieved significant success in many fields, the knowledge is encoded in a subsymbolic representation which is incompatible with symbolic systems such as planners. We propose Latplan, an unsupervised architecture combining deep learning and classical planning. Given only an unlabeled set of image pairs showing a subset of transitions allowed in the environment (training inputs), Latplan learns a complete propositional PDDL action model of the environment. Later, when a pair of images representing the initial and the goal states (planning inputs) is given, Latplan finds a plan to the goal state in a symbolic latent space and returns a visualized plan execution. We evaluate Latplan using image-based versions of 6 planning domains: 8-puzzle, 15-Puzzle, Blocksworld, Sokoban and Two variations of LightsOut.
翻译:古典规划者(Latplan)是一个不受监管的架构,将深层次学习和古典规划结合起来。Latplan只使用一组未贴标签的图像对应于显示环境允许的转变(培训投入)的一组图象,因此,Latplan学习了一套完整的PDDL行动环境模型。后来,当提供了一对代表初始和目标状态(规划投入)的图像时,Latplan在一个象征性的潜在空间找到一个目标状态的计划,并返回一个可视化的计划执行。我们用基于图像的6个规划领域:8个喷嘴、15个喷嘴、Blacksworld、Sokoban 和两个灯光输出变量来评估拉图计划。