Data augmentation (DA) has been widely investigated to facilitate model optimization in many tasks. However, in most cases, data augmentation is randomly performed for each training sample with a certain probability, which might incur content destruction and visual ambiguities. To eliminate this, in this paper, we propose an effective approach, dubbed SelectAugment, to select samples to be augmented in a deterministic and online manner based on the sample contents and the network training status. Specifically, in each batch, we first determine the augmentation ratio, and then decide whether to augment each training sample under this ratio. We model this process as a two-step Markov decision process and adopt Hierarchical Reinforcement Learning (HRL) to learn the augmentation policy. In this way, the negative effects of the randomness in selecting samples to augment can be effectively alleviated and the effectiveness of DA is improved. Extensive experiments demonstrate that our proposed SelectAugment can be adapted upon numerous commonly used DA methods, e.g., Mixup, Cutmix, AutoAugment, etc, and improve their performance on multiple benchmark datasets of image classification and fine-grained image recognition.
翻译:对数据增强(DA)进行了广泛的调查,以便利在很多任务中实现模型优化。然而,在多数情况下,数据增强是随机地对每个培训样本进行,有一定的概率,可能会造成内容破坏和视觉模糊。为了消除这一点,我们在本文件中建议一种有效的方法,即称为选择选择增强,根据样本内容和网络培训状况,以确定和在线方式选择样本,加以扩大。具体地说,在每批中,我们首先确定增强率,然后决定是否根据这一比率增加每个培训样本。我们把这一过程作为两步马可夫决定程序进行模拟,并采用等级强化学习(HRL)来学习增强政策。这样,随机选择样本增加的消极影响可以有效减轻,并改进DA的效能。广泛的实验表明,我们提议的选择增强可以适应许多常用的DA方法,例如,Mixup、Cutmix、Automix、AutAugA等,并改进其在图像分类和精细图像识别的多重基准数据集方面的性。