Machine learning techniques are used in a wide range of domains. However, machine learning models often suffer from the problem of over-fitting. Many data augmentation methods have been proposed to tackle such a problem, and one of them is called mixup. Mixup is a recently proposed regularization procedure, which linearly interpolates a random pair of training examples. This regularization method works very well experimentally, but its theoretical guarantee is not adequately discussed. In this study, we aim to discover why mixup works well from the aspect of the statistical learning theory.
翻译:机器学习技术在广泛的领域使用,但机器学习模式往往受到过度装配问题的影响。许多数据增强方法被提出来应对这样一个问题,其中一个被称作混合。混合是一种最近提出的正规化程序,它从线性上将随机的一对培训实例套接在一起。这种正规化方法在实验中非常有效,但其理论保障没有得到充分讨论。在本研究中,我们的目标是从统计学习理论的方面了解为什么混合运作良好。