Reinforcement learning tasks are typically specified as Markov decision processes. This formalism has been highly successful, though specifications often couple the dynamics of the environment and the learning objective. This lack of modularity can complicate generalization of the task specification, as well as obfuscate connections between different task settings, such as episodic and continuing. In this work, we introduce the RL task formalism, that provides a unification through simple constructs including a generalization to transition-based discounting. Through a series of examples, we demonstrate the generality and utility of this formalism. Finally, we extend standard learning constructs, including Bellman operators, and extend some seminal theoretical results, including approximation errors bounds. Overall, we provide a well-understood and sound formalism on which to build theoretical results and simplify algorithm use and development.
翻译:强化学习任务通常以Markov决定程序为主。这种形式主义非常成功,尽管其规格往往将环境的动态和学习目标结合起来。这种缺乏模块化的做法可能会使任务规格的概括化复杂化,并混淆不同任务环境之间的联系,例如偶发性和连续性。在这项工作中,我们引入了RL任务格式主义,通过简单的构思提供统一,包括向基于过渡的贴现的概括化。我们通过一系列实例展示了这种形式主义的普遍性和实用性。最后,我们扩展了标准学习结构,包括Bellman操作者,并扩展了一些半理论性结果,包括近似误差界限。总的来说,我们提供了一种非常清楚和合理的形式主义,用以建立理论结果并简化算法的使用和发展。