The Mixup method (Zhang et al. 2018), which uses linearly interpolated data, has emerged as an effective data augmentation tool to improve generalization performance and the robustness to adversarial examples. The motivation is to curtail undesirable oscillations by its implicit model constraint to behave linearly at in-between observed data points and promote smoothness. In this work, we formally investigate this premise, propose a way to explicitly impose smoothness constraints, and extend it to incorporate with implicit model constraints. First, we derive a new function class composed of kernel-convoluted models (KCM) where the smoothness constraint is directly imposed by locally averaging the original functions with a kernel function. Second, we propose to incorporate the Mixup method into KCM to expand the domains of smoothness. In both cases of KCM and the KCM adapted with the Mixup, we provide risk analysis, respectively, under some conditions for kernels. We show that the upper bound of the excess risk is not slower than that of the original function class. The upper bound of the KCM with the Mixup remains dominated by that of the KCM if the perturbation of the Mixup vanishes faster than \(O(n^{-1/2})\) where \(n\) is a sample size. Using CIFAR-10 and CIFAR-100 datasets, our experiments demonstrate that the KCM with the Mixup outperforms the Mixup method in terms of generalization and robustness to adversarial examples.
翻译:混合法(Zhang等人,2018年)使用线性内插数据,它已成为一个有效的数据增强工具,用来改进一般性能和对角性实例的稳健性,目的是减少不受欢迎的振动,因为其隐含的模式限制使得在观察到的数据点之间线性地在观察到的数据点之间行事,并促进平稳。在这项工作中,我们正式调查这一前提,提出一种明确施加平稳限制的方法,并将它扩展到隐含的模式限制。首先,我们产生了一个新的功能类别,由内核混合模型(KCM)组成,其中平稳性限制直接由本地以内核功能平均原功能直接施加。第二,我们提议将混合方法纳入KCMMM,以扩大光滑度范围。在KCMM和KMMMM, 以较稳健的IMFR=1, 其MMCM=MR=l) 和MMM=R=CM=(KMNB) 其磁性比CMR=越快。