Real-world datasets commonly have noisy labels, which negatively affects the performance of deep neural networks (DNNs). In order to address this problem, we propose a label noise robust learning algorithm, in which the base classifier is trained on soft-labels that are produced according to a meta-objective. In each iteration, before conventional training, the meta-objective reshapes the loss function by changing soft-labels, so that resulting gradient updates would lead to model parameters with minimum loss on meta-data. Soft-labels are generated from extracted features of data instances, and the mapping function is learned by a single layer perceptron (SLP) network, which is called MetaLabelNet. Following, base classifier is trained by using these generated soft-labels. These iterations are repeated for each batch of training data. Our algorithm uses a small amount of clean data as meta-data, which can be obtained effortlessly for many cases. We perform extensive experiments on benchmark datasets with both synthetic and real-world noises. Results show that our approach outperforms existing baselines.
翻译:现实世界数据集通常都有噪音标签,对深神经网络(DNNs)的性能有负面影响。为了解决这一问题,我们提议采用标签噪音强的学习算法,让基级分类员在按照元目标制作的软标签上接受培训。在常规培训之前,元目标在每次迭代中都通过改变软标签来改变损失功能,从而导致的梯度更新会导致模型参数,在元数据上损失最小。软标签是从数据实例的提取特征中产生的,而绘图功能则通过一个称为MetaLabelNet的单层分立器(SLP)网络来学习。随后,基级分类员通过使用这些生成的软标签进行培训。对每批培训数据重复进行这些迭代。我们的算法使用少量清洁数据作为元数据,许多情况下可以不费力地获取这些数据。我们用合成和现实世界噪音对基准数据集进行了广泛的实验。结果显示,我们的方法超过了现有的基线。