This paper reviews exploration techniques in deep reinforcement learning. Exploration techniques are of primary importance when solving sparse reward problems. In sparse reward problems, the reward is rare, which means that the agent will not find the reward often by acting randomly. In such a scenario, it is challenging for reinforcement learning to learn rewards and actions association. Thus more sophisticated exploration methods need to be devised. This review provides a comprehensive overview of existing exploration approaches, which are categorized based on the key contributions as follows reward novel states, reward diverse behaviours, goal-based methods, probabilistic methods, imitation-based methods, safe exploration and random-based methods. Then, the unsolved challenges are discussed to provide valuable future research directions. Finally, the approaches of different categories are compared in terms of complexity, computational effort and overall performance.
翻译:本文审视了深层强化学习中的勘探技术。 探索技术在解决稀有的奖励问题时至关重要。 在稀少的奖励问题中,奖赏是罕见的,这意味着代理人不会经常通过随机行动来找到奖励。 在这样的情况下,对强化学习学习以学习奖赏和行动协会具有挑战性。 因此,需要设计更先进的勘探方法。 本次审查根据以下主要贡献对现有勘探方法进行了全面概述:奖励新国家、奖励不同行为、基于目标的方法、概率方法、仿照方法、安全勘探和随机方法。 然后,讨论尚未解决的挑战,以提供宝贵的未来研究方向。 最后,不同类别的方法在复杂性、计算努力和总体绩效方面进行了比较。