Deep neural networks need large amounts of labeled data to achieve good performance. In real-world applications, labels are usually collected from non-experts such as crowdsourcing to save cost and thus are noisy. In the past few years, deep learning methods for dealing with noisy labels have been developed, many of which are based on the small-loss criterion. However, there are few theoretical analyses to explain why these methods could learn well from noisy labels. In this paper, we theoretically explain why the widely-used small-loss criterion works. Based on the explanation, we reformalize the vanilla small-loss criterion to better tackle noisy labels. The experimental results verify our theoretical explanation and also demonstrate the effectiveness of the reformalization.
翻译:深神经网络需要大量贴有标签的数据才能取得良好的性能。 在现实世界的应用中,标签通常是从非专家那里收集的,如众包,以节省成本,因此是吵闹的。在过去几年里,已经开发了处理吵闹标签的深层学习方法,其中许多方法是以小损失标准为基础的。然而,很少有理论分析来解释为什么这些方法可以从吵闹标签中学习好。在本文中,我们理论上解释了广泛使用的小额损失标准为什么起作用。根据解释,将香草小额损失标准重新配置,以便更好地解决吵闹标签问题。实验结果证实了我们的理论解释,也证明了改革的有效性。