In adversarial machine learning, there was a common belief that robustness and accuracy hurt each other. The belief was challenged by recent studies where we can maintain the robustness and improve the accuracy. However, the other direction, whether we can keep the accuracy while improving the robustness, is conceptually and practically more interesting, since robust accuracy should be lower than standard accuracy for any model. In this paper, we show this direction is also promising. Firstly, we find even over-parameterized deep networks may still have insufficient model capacity, because adversarial training has an overwhelming smoothing effect. Secondly, given limited model capacity, we argue adversarial data should have unequal importance: geometrically speaking, a natural data point closer to/farther from the class boundary is less/more robust, and the corresponding adversarial data point should be assigned with larger/smaller weight. Finally, to implement the idea, we propose geometry-aware instance-reweighted adversarial training, where the weights are based on how difficult it is to attack a natural data point. Experiments show that our proposal boosts the robustness of standard adversarial training; combining two directions, we improve both robustness and accuracy of standard adversarial training.
翻译:在对立机器学习中,人们普遍认为,强性和准确性互为害处。最近的研究对信仰提出了挑战,我们可以保持强性和准确性。然而,另一个方向,即我们能否在提高强性的同时保持准确性,在概念上和实际上更加有趣,因为强性准确性应低于任何模型的标准准确性。在本文中,我们展示了这一方向也是有希望的。首先,我们发现甚至超分深度的深层网络可能仍然没有足够的模型能力,因为对立性培训具有压倒性的平滑效应。第二,由于模型能力有限,我们认为对立性数据具有不平等的重要性:从几何学角度讲,一个更接近/远于阶级边界的自然数据点比较弱/更强,相应的对立性数据点应当以更大/更轻的重量来分配。最后,我们提出这一想法,我们建议进行几何测量性对准度的对立性对准性对准性对准性培训,因为对准性培训的权重在于攻击自然数据点是多么困难。实验表明,我们的提议会增强标准的对立性培训的稳健性;将两个方向结合起来,我们改进了对立性培训的对立性。