Deep neural networks obtained by standard training have been constantly plagued by adversarial examples. Although adversarial training demonstrates its capability to defend against adversarial examples, unfortunately, it leads to an inevitable drop in the natural generalization. To address the issue, we decouple the natural generalization and the robust generalization from joint training and formulate different training strategies for each one. Specifically, instead of minimizing a global loss on the expectation over these two generalization errors, we propose a bi-expert framework called \emph{Generalist} where we simultaneously train base learners with task-aware strategies so that they can specialize in their own fields. The parameters of base learners are collected and combined to form a global learner at intervals during the training process. The global learner is then distributed to the base learners as initialized parameters for continued training. Theoretically, we prove that the risks of Generalist will get lower once the base learners are well trained. Extensive experiments verify the applicability of Generalist to achieve high accuracy on natural examples while maintaining considerable robustness to adversarial ones. Code is available at https://github.com/PKU-ML/Generalist.
翻译:在标准训练得到的深度神经网络经常受到对抗样例的困扰。虽然对抗训练证明了其防御对抗样例的能力,但不幸的是,它会导致自然泛化的不可避免下降。为了解决这个问题,我们将自然泛化和健壮泛化从联合训练中解耦,并为每种情况制定不同的训练策略。具体而言,我们提出了一种双专家框架,称为“通用主义”,其中我们同时使用面向任务的策略来训练基础学习器,使它们可以专注于各自的领域。在训练过程中,基础学习器的参数将被收集并组合成全局学习器。全局学习器随后分布给基础学习器作为持续训练的初始化参数。从理论上讲,我们证明了如果基础学习器训练得很好,通用主义的风险将降低。广泛的实验证明了通用主义的适用性,在保持相当鲁棒性的同时,在自然例子上实现了高准确性。代码可在https://github.com/PKU-ML/Generalist 上找到。