Image-mixing augmentations (e.g., Mixup and CutMix), which typically involve mixing two images, have become the de-facto training techniques for image classification. Despite their huge success in image classification, the number of images to be mixed has not been elucidated in the literature: only the naive K-image expansion has been shown to lead to performance degradation. This study derives a new K-image mixing augmentation based on the stick-breaking process under Dirichlet prior distribution. We demonstrate the superiority of our K-image expansion augmentation over conventional two-image mixing augmentation methods through extensive experiments and analyses: (1) more robust and generalized classifiers; (2) a more desirable loss landscape shape; (3) better adversarial robustness. Moreover, we show that our probabilistic model can measure the sample-wise uncertainty and boost the efficiency for network architecture search by achieving a 7-fold reduction in the search time. Code will be available at https://github.com/yjyoo3312/DCutMix-PyTorch.git.
翻译:图像混合扩增(例如Mixup和CutMix)已成为图像分类的事实训练技术。尽管它们在图像分类方面取得了巨大成功,但在文献中并没有阐述必须混合多少张图像:只有幼稚的K-image扩展被证明会导致性能下降。本研究基于Dirichlet先验分布推导了一种新的K-image混合扩增。我们通过广泛的实验和分析显示了我们的K-image扩展扩增法优于传统的两张图像混淆扩增法:(1)更强健和广义的分类器;(2)更理想的损失函数梯度;(3)更好的对抗鲁棒性。此外,我们展示了我们的概率模型可以测量每个样本的不确定性和通过实现7倍的搜索时间缩减来增强网络架构的搜索效率。代码可在https://github.com/yjyoo3312/DCutMix-PyTorch.git找到。