Regularization is essential for avoiding over-fitting to training data in network optimization, leading to better generalization of the trained networks. The label noise provides a strong implicit regularization by replacing the target ground truth labels of training examples by uniform random labels. However, it can cause undesirable misleading gradients due to the large loss associated with incorrect labels. We propose a first-order optimization method (Label-Noised Trim-SGD) that uses the label noise with the example trimming in order to remove the outliers based on the loss. The proposed algorithm is simple yet enables us to impose a large label-noise and obtain a better regularization effect than the original methods. The quantitative analysis is performed by comparing the behavior of the label noise, the example trimming, and the proposed algorithm. We also present empirical results that demonstrate the effectiveness of our algorithm using the major benchmarks and the fundamental networks, where our method has successfully outperformed the state-of-the-art optimization methods.
翻译:标签噪声通过统一随机标签取代培训范例的目标地面真实标签,从而提供了一种强烈的隐含的正规化效果。然而,它可能会因与错误标签有关的大量损失而造成不可取的误导梯度。我们建议了一种第一阶优化方法(Label-Noized Trim-SGD),使用标签噪声和标注的精细处理模型来去除基于损失的外部线。提议的算法很简单,但使我们能够实施一个大型标签噪音,并获得比原方法更好的正规化效果。定量分析是通过比较标签噪音、示例三角和拟议算法的行为进行的。我们还介绍了实验结果,这些结果显示了我们使用主要基准和基本网络算法的有效性,我们的方法成功地超过了最新最优化方法。