The combination of Reinforcement Learning (RL) with deep learning has led to a series of impressive feats, with many believing (deep) RL provides a path towards generally capable agents. However, the success of RL agents is often highly sensitive to design choices in the training process, which may require tedious and error-prone manual tuning. This makes it challenging to use RL for new problems, while also limits its full potential. In many other areas of machine learning, AutoML has shown it is possible to automate such design choices and has also yielded promising initial results when applied to RL. However, Automated Reinforcement Learning (AutoRL) involves not only standard applications of AutoML but also includes additional challenges unique to RL, that naturally produce a different set of methods. As such, AutoRL has been emerging as an important area of research in RL, providing promise in a variety of applications from RNA design to playing games such as Go. Given the diversity of methods and environments considered in RL, much of the research has been conducted in distinct subfields, ranging from meta-learning to evolution. In this survey we seek to unify the field of AutoRL, we provide a common taxonomy, discuss each area in detail and pose open problems which would be of interest to researchers going forward.
翻译:强化学习(RL)与深层次学习相结合,导致了一系列令人印象深刻的成就,许多相信(深)RL为通向一般能干的代理人提供了一条通向一般能的代理人的道路,然而,RL代理的成功往往对培训过程中的设计选择非常敏感,这可能需要烦琐和容易出错的手工调整。这使得使用RL解决新问题具有挑战性,同时也限制了它的全部潜力。在机器学习的许多其他领域,AutoML表明可以使这种设计选择自动化,并且在应用到RL时也取得了令人印象深刻的初步结果。然而,自动化强化学习(AutoRL)不仅涉及AutoML的标准应用,而且还包括了LAutoML特有的额外挑战,这自然产生了一套不同的方法。因此,AutoRL已经成为一个重要的研究领域,在从RNA设计到玩游戏(如Go)等的各种应用中提供了希望。鉴于RL公司所考虑的方法和环境的多样性,许多研究都是在不同的子领域进行的,从元学习到演进。在本次调查中,我们试图统一一个共同的会计领域。