Hierarchies of temporally decoupled policies present a promising approach for enabling structured exploration in complex long-term planning problems. To fully achieve this approach an end-to-end training paradigm is needed. However, training these multi-level policies has had limited success due to challenges arising from interactions between the goal-assigning and goal-achieving levels within a hierarchy. In this article, we consider the policy optimization process as a multi-agent process. This allows us to draw on connections between communication and cooperation in multi-agent RL, and demonstrate the benefits of increased cooperation between sub-policies on the training performance of the overall policy. We introduce a simple yet effective technique for inducing inter-level cooperation by modifying the objective function and subsequent gradients of higher-level policies. Experimental results on a wide variety of simulated robotics and traffic control tasks demonstrate that inducing cooperation results in stronger performing policies and increased sample efficiency on a set of difficult long time horizon tasks. We also find that goal-conditioned policies trained using our method display better transfer to new tasks, highlighting the benefits of our method in learning task-agnostic lower-level behaviors. Videos and code are available at: https://sites.google.com/berkeley.edu/cooperative-hrl.
翻译:暂时脱钩的政策等级制度为在复杂的长期规划问题中进行结构化探索提供了一个很有希望的方法,有助于在复杂的长期规划问题中进行结构化的探索。要完全实现这一方法,需要一种端对端的培训模式。然而,由于目标分配和实现目标层次之间在等级体系内的互动所产生的挑战,培训这些多层次的政策取得了有限的成功。在本条中,我们认为政策优化进程是一个多试办程序。这使我们能够利用多试剂RL的沟通与合作之间的联系,并展示了在总体政策培训绩效次级政策之间加强合作的好处。我们引入了一种简单而有效的技术,通过修改目标功能和随后更高层次政策的梯度来引导不同层次的合作。关于各种模拟机器人和交通控制任务的实验结果表明,在一系列困难的时间跨度任务中,促使合作更强有力地执行政策和提高抽样效率。我们还发现,利用我们的方法培训的有目标性的政策向新任务转移得更好,突出我们学习任务低层次行为的方法的好处。在https/developtionalgels。