We investigate the problem of designing optimal stealthy poisoning attacks on the control channel of Markov decision processes (MDPs). This research is motivated by the recent interest of the research community for adversarial and poisoning attacks applied to MDPs, and reinforcement learning (RL) methods. The policies resulting from these methods have been shown to be vulnerable to attacks perturbing the observations of the decision-maker. In such an attack, drawing inspiration from adversarial examples used in supervised learning, the amplitude of the adversarial perturbation is limited according to some norm, with the hope that this constraint will make the attack imperceptible. However, such constraints do not grant any level of undetectability and do not take into account the dynamic nature of the underlying Markov process. In this paper, we propose a new attack formulation, based on information-theoretical quantities, that considers the objective of minimizing the detectability of the attack as well as the performance of the controlled process. We analyze the trade-off between the efficiency of the attack and its detectability. We conclude with examples and numerical simulations illustrating this trade-off.
翻译:我们调查了设计对Markov决策程序控制渠道进行最佳隐性中毒袭击的问题。这一研究的动机是研究界最近对MDP应用的对抗性攻击和中毒攻击的兴趣以及强化学习方法。这些方法所产生的政策已经证明很容易受到干扰决策者观察结果的攻击。在这种攻击中,从监督学习中使用的对抗性例子中得到的启发,对对抗性扰动的振动根据某些规范受到限制,希望这种限制将使攻击无法察觉。然而,这种限制不会造成任何程度的不可察觉性,也没有考虑到Markov基本过程的动态性质。在本文中,我们根据信息理论数量,提出一种新的攻击性制剂,考虑尽量减少攻击的可探测性以及控制过程的进行。我们分析了攻击效率与可探测性之间的权衡。我们最后用实例和数字模拟来说明这一交易。