Adversarial training is arguably an effective but time-consuming way to train robust deep neural networks that can withstand strong adversarial attacks. As a response to the inefficiency, we propose the Dynamic Efficient Adversarial Training (DEAT), which gradually increases the adversarial iteration during training. Moreover, we theoretically reveal that the connection of the lower bound of Lipschitz constant of a given network and the magnitude of its partial derivative towards adversarial examples. Supported by this theoretical finding, we utilize the gradient's magnitude to quantify the effectiveness of adversarial training and determine the timing to adjust the training procedure. This magnitude based strategy is computational friendly and easy to implement. It is especially suited for DEAT and can also be transplanted into a wide range of adversarial training methods. Our post-investigation suggests that maintaining the quality of the training adversarial examples at a certain level is essential to achieve efficient adversarial training, which may shed some light on future studies.
翻译:反向培训可以说是一种有效但费时的方法,用于培训能够抵御强烈对抗性攻击的强大深层神经网络。作为对低效率的回应,我们建议采用动态高效反向培训(DEAT),逐步增加培训期间的对抗性迭代。此外,从理论上讲,我们发现,一个特定网络的下层连接及其部分衍生物与对抗性实例的大小。根据这一理论发现,我们利用梯度的大小来量化对抗性培训的有效性,并确定调整培训程序的时机。基于这一规模的战略在计算上是友好的,易于执行。它特别适合DEAT,也可以移植到广泛的对抗性培训方法中。我们的事后调查表明,保持某一层次的培训对抗性实例的质量对于实现有效的对抗性培训至关重要,这可能会为今后的研究提供一些启示。