We study universal consistency of non-i.i.d. processes in the context of online learning. A stochastic process is said to admit universal consistency if there exists a learner that achieves vanishing average loss for any measurable response function on this process. When the loss function is unbounded, Blanchard et al. showed that the only processes admitting strong universal consistency are those taking a finite number of values almost surely. However, when the loss function is bounded, the class of processes admitting strong universal consistency is much richer and its characterization could be dependent on the response setting (Hanneke). In this paper, we show that this class of processes is independent from the response setting thereby closing an open question (Hanneke, Open Problem 3). Specifically, we show that the class of processes that admit universal online learning is the same for binary classification as for multiclass classification with countable number of classes. Consequently, any output setting with bounded loss can be reduced to binary classification. Our reduction is constructive and practical. Indeed, we show that the nearest neighbor algorithm is transported by our construction. For binary classification on a process admitting strong universal learning, we prove that nearest neighbor successfully learns at least all finite unions of intervals.
翻译:我们研究的是在线学习中非i.i.d.进程的普遍一致性。据说,如果有一个学习者在这个过程中的任何可衡量的响应功能中实现了平均损失的消失,那么一个随机过程就承认了普遍一致性。当损失功能不受约束时,Blanchard 等人表明,唯一承认高度普遍一致性的程序是采用数量有限的数值的过程。然而,当损失功能受约束时,承认高度普遍一致性的流程类别就会更加丰富,其定性可能取决于响应设置(Hanneke)。在本文中,我们表明,这一类流程独立于响应设置,从而结束一个开放式问题(Hanneke, Open Ristle 3),我们表明,接受普遍在线学习的流程类别在二等分类方面与使用可点数的多级分类相同。因此,任何与受约束损失挂钩的输出都可简化为二进制分类。我们的削减是建设性和实用的。事实上,我们表明最近的邻居算法是通过我们的构建方式转换的。对于一个接受强度普遍学习的二进分类过程来说,我们证明最接近近的近一级的近一级的近一步学习。