Modern neural networks are over-parameterized and thus rely on strong regularization such as data augmentation and weight decay to reduce overfitting and improve generalization. The dominant form of data augmentation applies invariant transforms, where the learning target of a sample is invariant to the transform applied to that sample. We draw inspiration from human visual classification studies and propose generalizing augmentation with invariant transforms to soft augmentation where the learning target softens non-linearly as a function of the degree of the transform applied to the sample: e.g., more aggressive image crop augmentations produce less confident learning targets. We demonstrate that soft targets allow for more aggressive data augmentation, offer more robust performance boosts, work with other augmentation policies, and interestingly, produce better calibrated models (since they are trained to be less confident on aggressively cropped/occluded examples). Combined with existing aggressive augmentation strategies, soft target 1) doubles the top-1 accuracy boost across Cifar-10, Cifar-100, ImageNet-1K, and ImageNet-V2, 2) improves model occlusion performance by up to $4\times$, and 3) halves the expected calibration error (ECE). Finally, we show that soft augmentation generalizes to self-supervised classification tasks.
翻译:我们从人类视觉分类研究中汲取灵感,并提议通过不易变换到软增殖,学习目标非线性地将增殖普遍化,这是对抽样应用的变异程度的函数:例如,更积极的图像作物增殖产生更有信心的学习目标。 我们证明软目标允许更积极的数据扩增,提供更强大的性能增强,与其他增强政策合作,并令人感兴趣的是,产生更好的校准模型(因为经过培训,对激烈裁剪/隐蔽的例子不那么有信心)。 与现有的进取性增强战略相结合,软目标1,将整个Cifar-10、Cifar-100、图像Net-1K和图像网络-V2的上一级精度提升翻倍,使整个Cifar-10、Scifar-100、图像网络-1K和图像网络-V2的变异性化程度的变异性化程度提高两倍,使模型加固性化性能提高,达到4美元,软性增强性性增强性增强性能,与其他增强性政策合作,以及有趣的是,产生更好的校准模型(因为经过训练后,将预期的加固度错误减半)。