Mixup is a popular data augmentation technique for training deep neural networks where additional samples are generated by linearly interpolating pairs of inputs and their labels. This technique is known to improve the generalization performance in many learning paradigms and applications. In this work, we first analyze Mixup and show that it implicitly regularizes infinitely many directional derivatives of all orders. We then propose a new method to improve Mixup based on the novel insight. To demonstrate the effectiveness of the proposed method, we conduct experiments across various domains such as images, tabular data, speech, and graphs. Our results show that the proposed method improves Mixup across various datasets using a variety of architectures, for instance, exhibiting an improvement over Mixup by 0.8% in ImageNet top-1 accuracy.
翻译:混合是一种受欢迎的数据增强技术,用于培训深神经网络,在这些网络中,通过线性内插投入和标签生成更多的样本。这种技术已知可以改善许多学习模式和应用的通用性能。在这项工作中,我们首先分析混合,并表明它暗含地规范了所有订单的无限多方向衍生物。然后我们根据新颖的洞察力提出了改进混合的新方法。为了证明拟议方法的有效性,我们在图象、表格数据、语音和图表等不同领域进行实验。我们的结果显示,拟议的方法利用各种结构改进了各种数据集的混合,例如,在图像网上层-1精确度中,在组合上方-1精确度上方0.8%的基础上展示了0.8%的改进率。