Reinforcement learning in problems with symbolic state spaces is challenging due to the need for reasoning over long horizons. This paper presents a new approach that utilizes relational abstractions in conjunction with deep learning to learn a generalizable Q-function for such problems. The learned Q-function can be efficiently transferred to related problems that have different object names and object quantities, and thus, entirely different state spaces. We show that the learned generalized Q-function can be utilized for zero-shot transfer to related problems without an explicit, hand-coded curriculum. Empirical evaluations on a range of problems show that our method facilitates efficient zero-shot transfer of learned knowledge to much larger problem instances containing many objects.
翻译:在象征性国家空间存在问题时加强学习是困难的,因为需要从长远角度进行推理。本文件提出了一种新的方法,利用关系抽象,同时深思熟虑,学习这些问题的通用功能。所学的Q功能可以有效地转移到相关问题上,这些问题的物体名称和物体数量不同,因此,完全不同的国家空间。我们表明,所学的通用功能可以在没有明确的手工编码课程的情况下用于零发转至相关问题。对一系列问题的实证评估表明,我们的方法有助于将所学知识有效地零射向包含许多物体的更大问题实例。