Recent adversarial attack developments have made reinforcement learning more vulnerable, and different approaches exist to deploy attacks against it, where the key is how to choose the right timing of the attack. Some work tries to design an attack evaluation function to select critical points that will be attacked if the value is greater than a certain threshold. This approach makes it difficult to find the right place to deploy an attack without considering the long-term impact. In addition, there is a lack of appropriate indicators of assessment during attacks. To make the attacks more intelligent as well as to remedy the existing problems, we propose the reinforcement learning-based attacking framework by considering the effectiveness and stealthy spontaneously, while we also propose a new metric to evaluate the performance of the attack model in these two aspects. Experimental results show the effectiveness of our proposed model and the goodness of our proposed evaluation metric. Furthermore, we validate the transferability of the model, and also its robustness under the adversarial training.
翻译:最近的对抗性攻击发展使得加强学习更加脆弱,对它采用不同方法部署攻击,关键在于如何选择攻击的正确时机。有些工作试图设计攻击评价功能,以选择在价值大于某一阈值时将攻击的临界点。这种方法使得在不考虑长期影响的情况下很难找到部署攻击的适当地点。此外,在攻击期间缺乏适当的评估指标。为了使攻击更加聪明,并纠正现有问题,我们提议以学习为基础的加强攻击框架,方法是考虑攻击模式的有效性和自发性,同时我们还提出一个新的衡量标准,用以评价攻击模式在这两个方面的表现。实验结果表明我们拟议模式的有效性和我们拟议评价指标的良好性。此外,我们还验证了模型的可转移性,以及在对抗性训练下是否可靠。