Deep networks have achieved impressive results on a range of well-curated benchmark datasets. Surprisingly, their performance remains sensitive to perturbations that have little effect on human performance. In this work, we propose a novel extension of Mixup called Robustmix that regularizes networks to classify based on lower-frequency spatial features. We show that this type of regularization improves robustness on a range of benchmarks such as Imagenet-C and Stylized Imagenet. It adds little computational overhead and, furthermore, does not require a priori knowledge of a large set of image transformations. We find that this approach further complements recent advances in model architecture and data augmentation, attaining a state-of-the-art mCE of 44.8 with an EfficientNet-B8 model and RandAugment, which is a reduction of 16 mCE compared to the baseline.
翻译:深度网络在一系列经过良好策划的基准数据集上取得了令人印象深刻的结果。令人惊讶的是,它们的表现仍对几乎对人类表现无影响的扰动敏感。在本文中,我们提出了Mixup的一种新颖扩展,称为Robustmix,它规范网络,使其基于低频空间特征进行分类。我们展示了这种类型的正则化可以在一系列基准测试中提高鲁棒性,如Imagenet-C和Stylized Imagenet。它增加了很少的计算开销,而且不需要先验知识的大量图像变换。我们发现,这种方法进一步补充了模型架构和数据增广的最新进展,在EfficientNet-B8模型和RandAugment的情况下达到了44.8的最新mCE水平,与基线相比降低了16个mCE。