Deep neural networks are often not robust to semantically-irrelevant changes in the input. In this work we address the issue of robustness of state-of-the-art deep convolutional neural networks (CNNs) against commonly occurring distortions in the input such as photometric changes, or the addition of blur and noise. These changes in the input are often accounted for during training in the form of data augmentation. We have two major contributions: First, we propose a new regularization loss called feature-map augmentation (FMA) loss which can be used during finetuning to make a model robust to several distortions in the input. Second, we propose a new combined augmentations (CA) finetuning strategy, that results in a single model that is robust to several augmentation types at the same time in a data-efficient manner. We use the CA strategy to improve an existing state-of-the-art method called stability training (ST). Using CA, on an image classification task with distorted images, we achieve an accuracy improvement of on average 8.94% with FMA and 8.86% with ST absolute on CIFAR-10 and 8.04% with FMA and 8.27% with ST absolute on ImageNet, compared to 1.98% and 2.12%, respectively, with the well known data augmentation method, while keeping the clean baseline performance.
翻译:深心神经网络往往不健全,无法对投入进行与静脉相关的变化。 在这项工作中,我们处理的是最先进的深发神经网络(CNNs)的稳健性问题,以防止输入经常发生的扭曲,例如光度变化,或添加模糊和噪音。输入中的这些变化通常在培训中以数据增强的形式进行核算。我们有两个主要贡献:首先,我们提议了一个新的正规化损失,称为地貌图增强(FMA)损失,在微调期间可以使用,使模型对输入中的若干扭曲情况具有稳健性。第二,我们提出了一个新的联合增强神经网络(CA)微调战略,其结果为单一模式,对若干增强型投入通常发生扭曲,例如光度变化或添加模糊和噪音。我们使用CAA战略改进了一种叫作稳定培训的现有状态方法。我们有两个重要贡献:首先,我们利用CAA进行图像分类任务,我们实现了平均8.94%的准确性改进,使输入的模型达到8.86%的绝对值。第二,我们提出了一个新的合并增强(CA)的调整战略,在数据周期10和8.04 %的调整战略,以数据同时对若干种增强型增强型类型的增强型类型。