We consider the problem of multi-class classification, where a stream of adversarially chosen queries arrive and must be assigned a label online. Unlike traditional bounds which seek to minimize the misclassification rate, we minimize the total distance from each query to the region corresponding to its correct label. When the true labels are determined via a nearest neighbor partition -- i.e. the label of a point is given by which of $k$ centers it is closest to in Euclidean distance -- we show that one can achieve a loss that is independent of the total number of queries. We complement this result by showing that learning general convex sets requires an almost linear loss per query. Our results build off of regret guarantees for the geometric problem of contextual search. In addition, we develop a novel reduction technique from multiclass classification to binary classification which may be of independent interest.
翻译:我们考虑的是多级分类问题,在多级分类中,一连串的对立选择查询到达,必须在线分配标签。与试图尽量减少错误分类率的传统界限不同,我们尽量缩小每个查询与与其正确标签相对应的区域之间的总距离。当真正的标签通过最近的邻里分区确定时,即标出一个点的标签,用点的标签来表示它最接近于欧格利德距离的美元中心,我们表明,一个人可以实现一个与查询总数无关的损失。我们通过显示学习一般二次曲线组合需要每查询几乎线性损失来补充这一结果。我们的结果是在为背景搜索的几何问题提供遗憾保证的基础上形成的。此外,我们开发了一种从多级分类到二元分类的新式的削减技术,可能具有独立的兴趣。