Adversarial training, as one of the few certified defenses against adversarial attacks, can be quite complicated and time-consuming, while the results might not be robust enough. To address the issue of lack of robustness, ensemble methods were proposed, aiming to get the final output by weighting the selected results from repeatedly trained processes. It is proved to be very useful in achieving robust and accurate results, but the computational and memory costs are even higher. Snapshot ensemble, a new ensemble method that combines several local minima in a single training process to make the final prediction, was proposed recently, which reduces the time spent on training multiple networks and the memory to store the results. Based on the snapshot ensemble, we present a new method that is easier to implement: unlike original snapshot ensemble that seeks for local minima, our snapshot ensemble focuses on the last few iterations of a training and stores the sets of parameters from them. Our algorithm is much simpler but the results are no less accurate than the original ones: based on different hyperparameters and datasets, our snapshot ensemble has shown a 5% to 30% increase in accuracy when compared to the traditional adversarial training.
翻译:Adversarial 培训是针对对抗性攻击的少数经认证的防御手段之一,它可能是相当复杂和耗时的,尽管其结果可能不够强大。为解决缺乏稳健性的问题,提出了一整套方法,目的是通过权衡反复培训过程的选定结果来获得最后产出。事实证明,它对于取得稳健和准确的结果非常有用,但计算和记忆成本甚至更高。快照合奏是一个新的混合方法,它将一个单一的培训进程中的若干本地迷你结合在一起,以作出最后预测。最近提出了这种方法,减少了培训多个网络和存储结果的记忆所花的时间。根据快照合奏,我们提出了一个更容易执行的新方法:不同于最初的模拟合奏,我们的快照合奏侧重于培训的最后几处迭代,并储存来自它们的成套参数。我们的算术简单得多,但结果并不比原始方法准确得多:根据不同的超常参数和数据模型和存储结果,在将传统的30 % 模拟培训比为精确性地显示,我们的30 % 模拟培训比为精确性。