When many labels are possible, choosing a single one can lead to low precision. A common alternative, referred to as top-$K$ classification, is to choose some number $K$ (commonly around 5) and to return the $K$ labels with the highest scores. Unfortunately, for unambiguous cases, $K>1$ is too many and, for very ambiguous cases, $K \leq 5$ (for example) can be too small. An alternative sensible strategy is to use an adaptive approach in which the number of labels returned varies as a function of the computed ambiguity, but must average to some particular $K$ over all the samples. We denote this alternative average-$K$ classification. This paper formally characterizes the ambiguity profile when average-$K$ classification can achieve a lower error rate than a fixed top-$K$ classification. Moreover, it provides natural estimation procedures for both the fixed-size and the adaptive classifier and proves their consistency. Finally, it reports experiments on real-world image data sets revealing the benefit of average-$K$ classification over top-$K$ in practice. Overall, when the ambiguity is known precisely, average-$K$ is never worse than top-$K$, and, in our experiments, when it is estimated, this also holds.
翻译:当许多标签都有可能时,选择一个标签可以导致低精度。一个共同的替代方案,称为最高-K美元分类,是选择一些数字K$(通常在5美元左右),然后将美元标签退回最高分。不幸的是,对于明确的情况,K$>1是太多的,对于非常模糊的情况,5美元(例如,5美元)可能太小。另一个明智的替代战略是采用适应性方法,即返回的标签数量因计算模糊性而不同,但必须在所有样本中平均为某些特定K美元。我们指出这一替代的平均-K美元分类。本文正式说明了在平均-K美元分类能够达到比固定的最高-K美元分类低的错误率时的模糊性特征。此外,它为固定大小和适应性分类员提供了自然估计程序,并证明了其一致性。最后,它报告了对真实世界图像数据组的实验,显示平均-K美元分类比最高-K美元在实际中平均-K美元的好处。总体而言,当我们准确知道这种模糊性时,我们的平均-K美元是最高-K美元是最差的。