AutoAugment has sparked an interest in automated augmentation methods for deep learning models. These methods estimate image transformation policies for train data that improve generalization to test data. While recent papers evolved in the direction of decreasing policy search complexity, we show that those methods are not robust when applied to biased and noisy data. To overcome these limitations, we reformulate AutoAugment as a generalized automated dataset optimization (AutoDO) task that minimizes the distribution shift between test data and distorted train dataset. In our AutoDO model, we explicitly estimate a set of per-point hyperparameters to flexibly change distribution of train data. In particular, we include hyperparameters for augmentation, loss weights, and soft-labels that are jointly estimated using implicit differentiation. We develop a theoretical probabilistic interpretation of this framework using Fisher information and show that its complexity scales linearly with the dataset size. Our experiments on SVHN, CIFAR-10/100, and ImageNet classification show up to 9.3% improvement for biased datasets with label noise compared to prior methods and, importantly, up to 36.6% gain for underrepresented SVHN classes.
翻译:为克服这些限制,我们调整了自动调整,作为通用自动数据集优化(AutoDO)任务,以尽量减少测试数据和扭曲的火车数据集之间的分布变化。在我们的AutoDO模型中,我们明确估计了一套用于改进通用数据分布的每点超参数,以灵活改变列车数据的分布。特别是,我们包括了增益、减重和软标签的超参数,这些超参数是使用隐含差异共同估算的。我们利用Fisher公司的信息对这一框架进行理论概率化解释,并显示其复杂性与数据集大小的线性比例。我们在SVHN、CIFAR-10/100和图像网络分类方面的实验显示,与以往方法相比,带有标签噪音的偏差数据集改进率为9.3%,重要的是,代表SVHN类的增益高达36.6%。