We consider bounds on the generalization performance of the least-norm linear regressor, in the over-parameterized regime where it can interpolate the data. We describe a sense in which any generalization bound of a type that is commonly proved in statistical learning theory must sometimes be very loose when applied to analyze the least-norm interpolant. In particular, for a variety of natural joint distributions on training examples, any valid generalization bound that depends only on the output of the learning algorithm, the number of training examples, and the confidence parameter, and that satisfies a mild condition (substantially weaker than monotonicity in sample size), must sometimes be very loose -- it can be bounded below by a constant when the true excess risk goes to zero.
翻译:我们考虑过量参数化制度中最不中性线性递减者一般性能的界限,在这种制度下,它可以对数据进行内插。我们描述了一种感觉,即统计学习理论中通常证明的某类一般性约束在用于分析最低度线性递减者时有时必须是非常松散的。特别是,对于各种关于培训实例的自然联合分发,任何有效的一般性约束,仅取决于学习算法的产出、培训实例的数量和信心参数,以及满足一种温和条件(在样本大小上比单体体积弱得多)有时必须是非常松散的 -- -- 当真正的超重风险达到零时,它可以被一个常数约束在下面。