Deep learning has achieved excellent performance in various computer vision tasks, but requires a lot of training examples with clean labels. It is easy to collect a dataset with noisy labels, but such noise makes networks overfit seriously and accuracies drop dramatically. To address this problem, we propose an end-to-end framework called PENCIL, which can update both network parameters and label estimations as label distributions. PENCIL is independent of the backbone network structure and does not need an auxiliary clean dataset or prior information about noise, thus it is more general and robust than existing methods and is easy to apply. PENCIL can even be used repeatedly to obtain better performance. PENCIL outperforms previous state-of-the-art methods by large margins on both synthetic and real-world datasets with different noise types and noise rates. And PENCIL is also effective in multi-label classification tasks through adding a simple attention structure on backbone networks. Experiments show that PENCIL is robust on clean datasets, too.
翻译:深入学习在各种计算机愿景任务中取得了卓越的成绩,但需要大量使用清洁标签的培训实例。很容易收集带有噪音标签的数据集,但这种噪音使得网络过于贴近,大范围缩小。为了解决这个问题,我们提议了一个名为PENCIL的端对端框架,它可以更新网络参数和标签估计,作为标签分布。 PENCIL独立于主干网络结构,不需要辅助性清洁数据集或关于噪音的先前信息,因此它比现有方法更加笼统和有力,而且易于应用。 PENCIL甚至可以反复使用来获得更好的性能。 PENCIL在合成和现实世界数据集上大幅度地用不同的噪音类型和噪声率取代以前的先进方法。 PENCIL通过在主干网络上添加一个简单的关注结构,在多标签分类任务中也很有效。 实验显示, PENCIL在清洁的数据集上也很强大。