Adversarial training is promising for improving robustness of deep neural networks towards adversarial perturbations, especially on the classification task. The effect of this type of training on semantic segmentation, contrarily, just commences. We make the initial attempt to explore the defense strategy on semantic segmentation by formulating a general adversarial training procedure that can perform decently on both adversarial and clean samples. We propose a dynamic divide-and-conquer adversarial training (DDC-AT) strategy to enhance the defense effect, by setting additional branches in the target model during training, and dealing with pixels with diverse properties towards adversarial perturbation. Our dynamical division mechanism divides pixels into multiple branches automatically. Note all these additional branches can be abandoned during inference and thus leave no extra parameter and computation cost. Extensive experiments with various segmentation models are conducted on PASCAL VOC 2012 and Cityscapes datasets, in which DDC-AT yields satisfying performance under both white- and black-box attack.
翻译:Aversarial 培训有望提高深神经网络的稳健性,使其更有利于对抗性扰动,特别是对于分类任务而言。这种培训对语义分解的影响才刚刚开始。我们最初试图探索语义分解的防御战略,方法是制定一个一般的对抗性对立培训程序,在对抗性和清洁样本上都能以体面的方式发挥作用。我们提议了一种动态的分化对立培训(DDC-AT)战略,以加强防御效果,办法是在培训期间在目标模式中设置更多的分支,处理具有不同特性的像素,以对抗性交叉扰动。我们的动态分解机制将像素自动分为多个分支。注意到所有这些额外的分支在推断期间都可以放弃,因此没有额外的参数和计算成本。在PASCAL VOC 2012和Cowes数据集中进行了广泛的分解模型实验,在白箱和黑箱攻击下DDC-AT都取得了令人满意的性能。