Novel class discovery (NCD) aims at learning a model that transfers the common knowledge from a class-disjoint labelled dataset to another unlabelled dataset and discovers new classes (clusters) within it. Many methods have been proposed as well as elaborate training pipelines and appropriate objectives and considerably boosted the performance on NCD tasks. Despite all this, we find that the existing methods do not sufficiently take advantage of the essence of the NCD setting. To this end, in this paper, we propose to model both inter-class and intra-class constraints in NCD based on the symmetric Kullback-Leibler divergence (sKLD). Specifically, we propose an inter-class sKLD constraint to effectively exploit the disjoint relationship between labelled and unlabelled classes, enforcing the separability for different classes in the embedding space. In addition, we present an intra-class sKLD constraint to explicitly constrain the intra-relationship between samples and their augmentations and ensure the stability of the training process at the same time. We conduct extensive experiments on the popular CIFAR10, CIFAR100 and ImageNet benchmarks and successfully demonstrate that our method can establish a new state of the art and can achieve significantly performance improvements, e.g., 3.6\%/3.7\% clustering accuracy improvements on CIFAR100-50 dataset split under the task-aware/-agnostic evaluation protocol, over previous state-of-the-art methods.
翻译:创新类发现(NCD)旨在学习一种模式,将普通知识从等级分解的贴标签数据集转移到另一个未贴标签的数据集,并在其中发现新的类(群)。我们提出了许多方法,并详细制定了培训管道和适当目标,大大促进了NCD任务的业绩。尽管如此,我们发现现有方法没有充分利用NCD设置的本质。为此,我们提议根据对称的 Kullback-Leiperer 差异(sKLD),在NCD中建模跨级和内部的制约。具体地说,我们提出一个等级间SKLD限制,以有效利用贴标签和未贴标签的类之间的脱节关系,加强嵌入空间中不同类之间的分离。此外,我们提出了一种内部的SKLD限制,以明确限制样本及其放大之间的内部关系,并确保同一时间的培训进程的稳定。我们就通用的 CPFAR10、CIFAR100和图像网络基准进行广泛的实验,并成功展示了SKLM-M-CS-366号任务下的业绩改进。我们的方法可以确定以往的进度。