Reinforcement learning allows machines to learn from their own experience. Nowadays, it is used in safety-critical applications, such as autonomous driving, despite being vulnerable to attacks carefully crafted to either prevent that the reinforcement learning algorithm learns an effective and reliable policy, or to induce the trained agent to make a wrong decision. The literature about the security of reinforcement learning is rapidly growing, and some surveys have been proposed to shed light on this field. However, their categorizations are insufficient for choosing an appropriate defense given the kind of system at hand. In our survey, we do not only overcome this limitation by considering a different perspective, but we also discuss the applicability of state-of-the-art attacks and defenses when reinforcement learning algorithms are used in the context of autonomous driving.
翻译:强化学习使机器能够从自身的经验中学习。 如今,它被用于安全关键应用,例如自主驾驶,尽管它很容易受到精心设计的攻击,以防止强化学习算法学习有效和可靠的政策,或促使受过训练的代理人作出错误的决定。关于强化学习安全性的文献正在迅速增加,并提议进行一些调查,以揭示这方面的情况。然而,鉴于现有系统的类型,它们的分类不足以选择适当的防御。在我们的调查中,我们不仅从不同的角度来克服这一限制,而且我们还讨论了在自主驾驶中使用强化学习算法时,最先进的攻击和防御的适用性。