Heuristic search algorithms, e.g. A*, are the commonly used tools for pathfinding on grids, i.e. graphs of regular structure that are widely employed to represent environments in robotics, video games etc. Instance-independent heuristics for grid graphs, e.g. Manhattan distance, do not take the obstacles into account and, thus, the search led by such heuristics performs poorly in the obstacle-rich environments. To this end, we suggest learning the instance-dependent heuristic proxies that are supposed to notably increase the efficiency of the search. The first heuristic proxy we suggest to learn is the correction factor, i.e. the ratio between the instance independent cost-to-go estimate and the perfect one (computed offline at the training phase). Unlike learning the absolute values of the cost-to-go heuristic function, which was known before, when learning the correction factor the knowledge of the instance-independent heuristic is utilized. The second heuristic proxy is the path probability, which indicates how likely the grid cell is lying on the shortest path. This heuristic can be utilized in the Focal Search framework as the secondary heuristic, allowing us to preserve the guarantees on the bounded sub-optimality of the solution. We learn both suggested heuristics in a supervised fashion with the state-of-the-art neural networks containing attention blocks (transformers). We conduct a thorough empirical evaluation on a comprehensive dataset of planning tasks, showing that the suggested techniques i) reduce the computational effort of the A* up to a factor of $4$x while producing the solutions, which costs exceed the costs of the optimal solutions by less than $0.3$% on average; ii) outperform the competitors, which include the conventional techniques from the heuristic search, i.e. weighted A*, as well as the state-of-the-art learnable planners.
翻译:超常搜索算法, 例如 A*, 是用来在网格上进行路由调查的常用工具, 即常态结构图, 被广泛用于代表机器人、 视频游戏等环境中的环境。 曼哈顿距离等网格图中, 不考虑障碍, 因此, 由这种超常学引导的搜索在障碍丰富的环境中表现不佳。 为此, 我们建议学习基于实例的超常预估工具, 以显著提高搜索效率。 我们建议学习的第一个超常结构图是校正系数, 即独立成本对go估计与完美图( 曼哈顿距离, 不考虑曼哈顿距离等网格图中独立的超常态图象数的比重。 不同于学习成本对古代数功能的绝对值, 在学习校正系数时, 亚超常值的代数是路径概率, 这表示电网格组在最短的时, 将成本对亚异性计算方法进行最短路段的比值。