Weakly supervised multi-label classification (WSML) task, which is to learn a multi-label classification using partially observed labels per image, is becoming increasingly important due to its huge annotation cost. In this work, we first regard unobserved labels as negative labels, casting the WSML task into noisy multi-label classification. From this point of view, we empirically observe that memorization effect, which was first discovered in a noisy multi-class setting, also occurs in a multi-label setting. That is, the model first learns the representation of clean labels, and then starts memorizing noisy labels. Based on this finding, we propose novel methods for WSML which reject or correct the large loss samples to prevent model from memorizing the noisy label. Without heavy and complex components, our proposed methods outperform previous state-of-the-art WSML methods on several partial label settings including Pascal VOC 2012, MS COCO, NUSWIDE, CUB, and OpenImages V3 datasets. Various analysis also show that our methodology actually works well, validating that treating large loss properly matters in a weakly supervised multi-label classification. Our code is available at https://github.com/snucml/LargeLossMatters.
翻译:微弱监管的多标签分类(WSML)任务,即学习使用每个图像部分观察的标签进行多标签分类,由于成本巨大,其重要性日益增大。在这项工作中,我们首先将未观察到的标签视为负标签,将WSML任务转化为吵闹的多标签分类。从这个角度看,我们从经验上观察到,最初在吵闹的多级环境中发现的记忆化效应也发生在多标签设置中。这就是,模型首先学习清洁标签的表述,然后开始记忆噪音标签。基于这一发现,我们为WSML提出了新颖的方法,拒绝或纠正大损失样本,以防止将大号标签混杂音。没有重和复杂的部件,我们提出的方法在包括Pascal VOC 2012、 MS CO、 NUSWIDE、 CUB 和 OpenIMages V3数据集在内的几个部分标签设置中超越了先前的状态-艺术 WSML方法。 VSM 分析还表明,我们的方法实际上运行薄弱/MLO。