Mixup-based data augmentations have achieved great success as regularizers for deep neural networks. However, existing methods rely on deliberately handcrafted mixup policies, which ignore or oversell the semantic matching between mixed samples and labels. Driven by their prior assumptions, early methods attempt to smooth decision boundaries by random linear interpolation while others focus on maximizing class-related information via offline saliency optimization. As a result, the issue of label mismatch has not been well addressed. Additionally, the optimization stability of mixup training is constantly troubled by the label mismatch. To address these challenges, we first reformulate mixup for supervised classification as two sub-tasks, mixup sample generation and classification, then propose Automatic Mixup (AutoMix), a revolutionary mixup framework. Specifically, a learnable lightweight Mix Block (MB) with a cross-attention mechanism is proposed to generate a mixed sample by modeling a fair relationship between the pair of samples under direct supervision of the corresponding mixed label. Moreover, the proposed Momentum Pipeline (MP) enhances training stability and accelerates convergence on top of making the Mix Block fully trained end-to-end. Extensive experiments on five popular classification benchmarks show that the proposed approach consistently outperforms leading methods by a large margin.
翻译:作为深层神经网络的正规化者,基于混合的数据增强工作取得了巨大成功,但是,现有方法依靠的是有意手工制造的混合政策,这些方法忽视或过度销售混合样品和标签之间的语义匹配。由先前的假设驱使,早期方法试图通过随机线性内插来平滑决定界限,而其他方法则侧重于通过离线显著优化最大限度地增加与阶级有关的信息。因此,标签错配问题没有很好地得到解决。此外,混合培训的最优化稳定性不断受到标签错配的困扰。为了应对这些挑战,我们首先重新配置组合,作为两个子任务进行监管的分类,混合样品生成和分类,然后提出自动混合(AutomiMix),即革命性混合框架。具体地说,建议采用一个具有交叉注意机制的可学习的轻质混合区块(MB),通过在相应混合标签的直接监管下对一对一对样品进行建模。此外,拟议的Momentum Pippeline(MP)将加强培训稳定性,并加速在最高级的Mix-Block 方法上,通过经过全面培训的五级最终的模型显示完整的大比例分析方法,在最高级的模型上进行最上展示。