Mixup is a popular data-dependent augmentation technique for deep neural networks, which contains two sub-tasks, mixup generation and classification. The community typically confines mixup to supervised learning (SL) and the objective of generation sub-task is fixed to the sampled pairs instead of considering the whole data manifold. To overcome such limitations, we systematically study the objectives of two sub-tasks and propose Scenario-Agostic Mixup for both SL and Self-supervised Learning (SSL) scenarios, named SAMix. Specifically, we hypothesize and verify the core objective of mixup generation as optimizing the local smoothness between two classes subject to global discrimination from other classes. Based on this discovery, $\eta$-balanced mixup loss is proposed for complementary training of the two sub-tasks. Meanwhile, the generation sub-task is parameterized as an optimizable module, Mixer, which utilizes an attention mechanism to generate mixed samples without label dependency. Extensive experiments on SL and SSL tasks demonstrate that SAMix consistently outperforms leading methods by a large margin.
翻译:为了克服这些局限性,我们系统地研究两个子任务的目标,并为SL和自监学习(SSS)设想方案提出情景-优异混合方法,称为SAMix。具体地说,我们虚度和核实混合产生的核心目标,以优化受其他类别全球歧视的两个班级之间的局部平稳状态。基于这一发现,提议为两个子任务的补充培训提供美元/元平衡混合损失。与此同时,生成子任务作为可选模块Mixer的参数,该模块使用关注机制生成混合样本,而无需贴标签。关于SAMix和SSL任务的广泛实验表明,SAMix始终以大幅度的方式超越领先方法。