Data augmentation is widely known as a simple yet surprisingly effective technique for regularizing deep networks. Conventional data augmentation schemes, e.g., flipping, translation or rotation, are low-level, data-independent and class-agnostic operations, leading to limited diversity for augmented samples. To this end, we propose a novel semantic data augmentation algorithm to complement traditional approaches. The proposed method is inspired by the intriguing property that deep networks are effective in learning linearized features, i.e., certain directions in the deep feature space correspond to meaningful semantic transformations, e.g., changing the background or view angle of an object. Based on this observation, translating training samples along many such directions in the feature space can effectively augment the dataset for more diversity. To implement this idea, we first introduce a sampling based method to obtain semantically meaningful directions efficiently. Then, an upper bound of the expected cross-entropy (CE) loss on the augmented training set is derived by assuming the number of augmented samples goes to infinity, yielding a highly efficient algorithm. In fact, we show that the proposed implicit semantic data augmentation (ISDA) algorithm amounts to minimizing a novel robust CE loss, which adds minimal extra computational cost to a normal training procedure. In addition to supervised learning, ISDA can be applied to semi-supervised learning tasks under the consistency regularization framework, where ISDA amounts to minimizing the upper bound of the expected KL-divergence between the augmented features and the original features. Although being simple, ISDA consistently improves the generalization performance of popular deep models (e.g., ResNets and DenseNets) on a variety of datasets, i.e., CIFAR-10, CIFAR-100, SVHN, ImageNet, and Cityscapes.
翻译:常规数据增强计划,例如翻转、翻译或旋转,是低层次的、数据独立和等级的操作,导致增加样本的多样性有限。为此,我们提出一个新的语义数据增强算法,以补充传统方法。拟议方法的灵感来自深层网络在学习线性特征方面行之有效的令人感兴趣的属性,即深层网络的某些方向与有意义的直线性能转变相对应,例如,翻转、翻译或旋转等。常规数据增强计划是低层次的,根据这种观察,将功能空间中许多此类方向的培训样本转换成高层次的,可以有效地增加数据的多样性。为了落实这一想法,我们首先采用基于取样的方法,以便有效地获得具有实际意义的方向。然后,在强化的训练中,通过假设增强的样本数量为直线性、精度的原始变异异性、高度的算法。事实上,我们显示,在深度的精度变异性能中,将预隐含的机能变异性能转化为常规数据变异性化过程。