In this paper, we study the problem of learning image classification models in the presence of label noise. We revisit a simple compression regularization named Nested Dropout. We find that Nested Dropout, though originally proposed to perform fast information retrieval and adaptive data compression, can properly regularize a neural network to combat label noise. Moreover, owing to its simplicity, it can be easily combined with Co-teaching to further boost the performance. Our final model remains simple yet effective: it achieves comparable or even better performance than the state-of-the-art approaches on two real-world datasets with label noise which are Clothing1M and ANIMAL-10N. On Clothing1M, our approach obtains 74.9% accuracy which is slightly better than that of DivideMix. On ANIMAL-10N, we achieve 84.1% accuracy while the best public result by PLC is 83.4%. We hope that our simple approach can be served as a strong baseline for learning with label noise. Our implementation is available at https://github.com/yingyichen-cyy/Nested-Co-teaching.
翻译:在本文中,我们研究了在标签噪音面前学习图像分类模型的问题。我们重新审视了名为Nesed Dropout的简单压缩规范化问题。我们发现,虽然最初建议快速信息检索和适应性数据压缩,但Nested 辍学可以适当地规范神经网络,以对抗标签噪音。此外,由于其简单化,它可以很容易地与共同教学相结合,以进一步提升性能。我们的最后模型仍然简单而有效:它比在两个真实世界数据集上采用最先进的方法,即Stragy1M和ANINAAL-10N。在服装1M中,我们的方法获得了74.9%的精度,比ExceptMix略好一些。在AMINAL-10N上,我们实现了84.1%的精度,而PLC的最佳公共结果为83.4%。我们希望我们简单的方法能够成为用标签噪音学习的强有力基线。我们的实施方法可以在https://github.com/yyyichen-cyyy/Nest-Coachinginginginginginging。