In this paper, we study the partial multi-label (PML) image classification problem, where each image is annotated with a candidate label set consists of multiple relevant labels and other noisy labels. Existing PML methods typically design a disambiguation strategy to filter out noisy labels by utilizing prior knowledge with extra assumptions, which unfortunately is unavailable in many real tasks. Furthermore, because the objective function for disambiguation is usually elaborately designed on the whole training set, it can be hardly optimized in a deep model with SGD on mini-batches. In this paper, for the first time we propose a deep model for PML to enhance the representation and discrimination ability. On one hand, we propose a novel curriculum based disambiguation strategy to progressively identify ground-truth labels by incorporating the varied difficulties of different classes. On the other hand, a consistency regularization is introduced for model retraining to balance fitting identified easy labels and exploiting potential relevant labels. Extensive experimental results on the commonly used benchmark datasets show the proposed method significantly outperforms the SOTA methods.
翻译:在本文中,我们研究了部分多标签(PML)图像分类问题,每张图象都配有候选标签,由多个相关标签和其他吵闹标签组成。现有的PML方法通常设计一种模糊化战略,利用先前的知识,加上额外假设来过滤噪音标签,但不幸的是,在许多实际任务中这些假设是无法做到的。此外,由于脱混的目标功能通常是在整个培训中精心设计的,因此很难在与小型篮子上的SGD的深层模型中加以优化。在本文中,我们首次为PML提出了一种深层模型,以提高代表性和歧视能力。一方面,我们提出了基于模糊化的新课程战略,以通过纳入不同类别的各种困难,逐步确定地标。另一方面,对模型再培训采用了一致性规范,以平衡已查明的简单标签和利用潜在相关标签。在常用的基准数据集上的广泛实验结果显示拟议方法大大优于SOTA方法。