Ensuring safety of reinforcement learning (RL) algorithms is crucial for many real-world tasks. However, vanilla RL does not guarantee safety for an agent. In recent years, several methods have been proposed to provide safety guarantees for RL. To the best of our knowledge, there is no comprehensive comparison of these provably safe RL methods. We therefore introduce a categorization for existing provably safe RL methods, and present the theoretical foundations for both continuous and discrete action spaces. Additionally, we evaluate provably safe RL on an inverted pendulum. In the experiments, it is shown that indeed only provably safe RL methods guarantee safety.
翻译:确保强化学习(RL)算法的安全性对于许多现实世界的任务至关重要,然而,香草RL并不能保证代理人的安全性。近些年来,提出了为RL提供安全保障的若干方法。据我们所知,这些安全性RL方法没有全面比较。因此,我们为现有的安全性RL方法进行分类,并为连续和离散的动作空间提供理论基础。此外,我们评估了倒置的钟摆上安全性RL。实验表明,确实只有安全性RL方法才能保证安全性。