Multi-label learning in the presence of missing labels (MLML) is a challenging problem. Existing methods mainly focus on the design of network structures or training schemes, which increase the complexity of implementation. This work seeks to fulfill the potential of loss function in MLML without increasing the procedure and complexity. Toward this end, we propose two simple yet effective methods via robust loss design based on an observation that a model can identify missing labels during training with a high precision. The first is a novel robust loss for negatives, namely the Hill loss, which re-weights negatives in the shape of a hill to alleviate the effect of false negatives. The second is a self-paced loss correction (SPLC) method, which uses a loss derived from the maximum likelihood criterion under an approximate distribution of missing labels. Comprehensive experiments on a vast range of multi-label image classification datasets demonstrate that our methods can remarkably boost the performance of MLML and achieve new state-of-the-art loss functions in MLML.
翻译:在缺少标签的情况下,多标签学习是一个具有挑战性的问题。现有方法主要侧重于网络结构或培训计划的设计,这增加了执行的复杂性。这项工作力求在不增加程序和复杂性的情况下实现MLML中损失功能的潜力。为此,我们提出两种简单而有效的方法,其依据是观察到一个模型可以在培训期间以高精确度识别缺失标签,通过稳健的损失设计,提出两种简单而有效的方法。第一种是对负值,即山值损失进行新的强力损失,即以山值为形状的负值重新加权负值,以减轻虚假负值的影响。第二种是自行测速的损失纠正方法,这种方法使用根据大致缺失标签分布的最大可能性标准产生的损失。关于广泛多标签图像分类数据集的全面实验表明,我们的方法可以显著提高MLMLML的性能,并在MLML中实现新的状态损失功能。