In modern classification tasks, the number of labels is getting larger and larger, as is the size of the datasets encountered in practice. As the number of classes increases, class ambiguity and class imbalance become more and more problematic to achieve high top-1 accuracy. Meanwhile, Top-K metrics (metrics allowing K guesses) have become popular, especially for performance reporting. Yet, proposing top-K losses tailored for deep learning remains a challenge, both theoretically and practically. In this paper we introduce a stochastic top-K hinge loss inspired by recent developments on top-K calibrated losses. Our proposal is based on the smoothing of the top-K operator building on the flexible "perturbed optimizer" framework. We show that our loss function performs very well in the case of balanced datasets, while benefiting from a significantly lower computational time than the state-of-the-art top-K loss function. In addition, we propose a simple variant of our loss for the imbalanced case. Experiments on a heavy-tailed dataset show that our loss function significantly outperforms other baseline loss functions.
翻译:在现代分类任务中,标签数量越来越多,数量也越来越多,实际中遇到的数据集规模也越来越多。随着班级数量的增加,班级的模糊性和阶级不平衡性越来越成问题,以达到最高一级至一级准确性。与此同时,高K度指标(允许K猜测的度量)已经很受欢迎,特别是在业绩报告方面。然而,提出用于深层次学习的高K级损失在理论上和实际上仍然是一个挑战。在本文件中,我们引入了一种由高K级校准损失最新动态所启发的随机顶级链条损失。我们的建议基于在灵活的“周遭优化”框架上平滑的顶K级操作员。我们表明,在平衡的数据集中,我们的损失功能表现得非常好,同时得益于比最高级K级损失功能低得多的计算时间。此外,我们为不平衡的个案提出了一个我们损失损失的简单变式。在重的数据集上进行的实验表明,我们的损失功能大大优于其他基线损失功能。