Mix-up training approaches have proven to be effective in improving the generalization ability of Deep Neural Networks. Over the years, the research community expands mix-up methods into two directions, with extensive efforts to improve saliency-guided procedures but minimal focus on the arbitrary path, leaving the randomization domain unexplored. In this paper, inspired by the superior qualities of each direction over one another, we introduce a novel method that lies at the junction of the two routes. By combining the best elements of randomness and saliency utilization, our method balances speed, simplicity, and accuracy. We name our method R-Mix following the concept of "Random Mix-up". We demonstrate its effectiveness in generalization, weakly supervised object localization, calibration, and robustness to adversarial attacks. Finally, in order to address the question of whether there exists a better decision protocol, we train a Reinforcement Learning agent that decides the mix-up policies based on the classifier's performance, reducing dependency on human-designed objectives and hyperparameter tuning. Extensive experiments further show that the agent is capable of performing at the cutting-edge level, laying the foundation for a fully automatic mix-up. Our code is released at [https://github.com/minhlong94/Random-Mixup].
翻译:事实证明,混合培训方法在提高深神经网络的普及能力方面是有效的。多年来,研究界将混合方法扩展为两个方向,广泛努力改进显著引导程序,但很少关注任意路径,使随机化域没有探索。在本文中,由于每个方向的优异性,我们引入了一种新颖的方法,它位于两条路径的交汇点。通过将随机性和突出利用的最佳要素、我们的方法平衡速度、简单性和准确性结合起来,我们用“兰多姆混合”的概念命名了我们的R-Mix方法。我们展示了该方法在一般化、低监管对象定位、校准和强度以适应对抗性攻击方面的有效性。最后,为了解决是否有更好的决策协议的问题,我们培训了一种强化学习剂,根据分级器的性能决定混合政策,减少对人类设计目标的依赖性和超准度调整。广泛的实验进一步表明,该代理人有能力在截断层/混合水平上执行[马吉姆·马克斯]的代码,为完全的自动建立基础。