Deep hierarchical reinforcement learning has gained a lot of attention in recent years due to its ability to produce state-of-the-art results in challenging environments where non-hierarchical frameworks fail to learn useful policies. However, as problem domains become more complex, deep hierarchical reinforcement learning can become inefficient, leading to longer convergence times and poor performance. We introduce the Deep Nested Agent framework, which is a variant of deep hierarchical reinforcement learning where information from the main agent is propagated to the low level $nested$ agent by incorporating this information into the nested agent's state. We demonstrate the effectiveness and performance of the Deep Nested Agent framework by applying it to three scenarios in Minecraft with comparisons to a deep non-hierarchical single agent framework, as well as, a deep hierarchical framework.
翻译:近些年来,深层次强化学习由于能够在非等级框架无法学习有用政策的挑战环境中产生最新成果而引起很多关注。然而,随着问题领域变得更加复杂,深层次强化学习可能变得效率低下,导致趋同时间延长和业绩不佳。我们引入了深层层强化学习框架,这是深层层强化学习的一种变体,通过将这些信息纳入嵌巢剂的状态,将主要代理商的信息传播到低水平的一美元代理商。我们通过将其应用于采矿工业中的三种情景,与深层非等级单一代理商框架进行比较,以及深层次的等级框架,展示了深层内层剂框架的有效性和绩效。