We study the close interplay between error and compression in the non-parametric multiclass classification setting in terms of prototype learning rules. We focus in particular on a close variant of a recently proposed compression-based learning rule termed OptiNet. Beyond its computational merits, this rule has been recently shown to be universally consistent in any metric instance space that admits a universally consistent rule -- the first learning algorithm known to enjoy this property. However, its error and compression rates have been left open. Here we derive such rates in the case where instances reside in Euclidean space under commonly posed smoothness and tail conditions on the data distribution. We first show that OptiNet achieves non-trivial compression rates while enjoying near minimax-optimal error rates. We then proceed to study a novel general compression scheme for further compressing prototype rules that locally adapts to the noise level without sacrificing accuracy. Applying it to OptiNet, we show that under a geometric margin condition, further gain in the compression rate is achieved. Experimental results comparing the performance of the various methods are presented.
翻译:我们从原型学习规则的角度研究非参数性多级分类设置的错误和压缩之间的密切相互作用。我们特别侧重于最近提议的压缩学习规则OptiNet的近似变体。除了计算优点外,这项规则最近在任何允许普遍一致规则(已知享有这一属性的首个学习算法)的计量空间中被证明是普遍一致的。然而,它的错误和压缩率一直没有被打开。我们在这里得出这样的比率,如果在Euclidean空间的情况中,数据分布通常呈现平滑和尾端条件。我们首先显示,OptiNet在享受接近微量最大最佳误差率的同时,实现了非三角压缩率。我们接着开始研究一个新的一般压缩办法,进一步压缩当地适应噪音水平的原型规则,同时不牺牲准确性。我们将其应用到OptiNet,我们表明,在几何边距条件下,压缩率将进一步提高。比较各种方法的性能的实验结果将会得到体现。