In the context of adversarial robustness, a single model does not usually have enough power to defend against all possible adversarial attacks, and as a result, has sub-optimal robustness. Consequently, an emerging line of work has focused on learning an ensemble of neural networks to defend against adversarial attacks. In this work, we take a principled approach towards building robust ensembles. We view this problem from the perspective of margin-boosting and develop an algorithm for learning an ensemble with maximum margin. Through extensive empirical evaluation on benchmark datasets, we show that our algorithm not only outperforms existing ensembling techniques, but also large models trained in an end-to-end fashion. An important byproduct of our work is a margin-maximizing cross-entropy (MCE) loss, which is a better alternative to the standard cross-entropy (CE) loss. Empirically, we show that replacing the CE loss in state-of-the-art adversarial training techniques with our MCE loss leads to significant performance improvement.
翻译:在对抗性强力的背景下,单一模型通常没有足够的力量来抵御所有可能的对抗性攻击,因此,它具有亚于最佳的强力。因此,正在形成的一行工作侧重于学习一整套神经网络来抵御对抗性攻击。在这项工作中,我们采取有原则的方法来建立强力的组合。我们从边际促进的角度看待这一问题,并开发一种算法来学习一种具有最大幅度的共性。通过对基准数据集的广泛经验性评价,我们表明我们的算法不仅优于现有的混合技术,而且优于以端到端方式培训的大型模型。我们工作的一个重要副产品是使跨作物损失的边际最大化,这是标准跨作物损失的更好替代。我们现时表明,用我们的MCE损失取代了最新水平的对抗性训练技术的CE损失。