The phenomenon of benign overfitting, where a predictor perfectly fits noisy training data while attaining low expected loss, has received much attention in recent years, but still remains not fully understood beyond simple linear regression setups. In this paper, we show that for regression, benign overfitting is ``biased'' towards certain types of problems, in the sense that its existence on one learning problem precludes its existence on other learning problems. On the negative side, we use this to argue that one should not expect benign overfitting to occur in general, for several natural extensions of the plain linear regression problems studied so far. We then turn to classification problems, and show that the situation there is much more favorable. Specifically, we consider a model where an arbitrary input distribution of some fixed dimension $k$ is concatenated with a high-dimensional distribution, and prove that the max-margin predictor (to which gradient-based methods are known to converge in direction) is asymptotically biased towards minimizing the expected \emph{squared hinge loss} w.r.t. the $k$-dimensional distribution. This allows us to reduce the question of benign overfitting in classification to the simpler question of whether this loss is a good surrogate for the misclassification error, and use it to show benign overfitting in some new settings.
翻译:良性超配现象 — 良性超配现象 — — 预言者完全适合杂乱的培训数据,同时又能达到预期的低损失水平 — — 近些年来受到了很多关注,但除了简单的线性回归设置之外,仍然没有得到完全理解。 在本文中,我们显示,对于回归而言,良性超配是对某些类型的问题的“偏差 ”, 因为它存在于一个学习问题中,因此它无法在其他学习问题中存在。 在负面方面,我们用它来论证,对于迄今为止研究的简单线性回归问题的若干自然延伸,人们不应期望出现良性超配。 我们接着讨论分类问题,并表明那里的情况更加有利。 具体地说,我们考虑了一种模式,即某些固定尺寸的硬化输入分配是“ $k$k$ ”, 与高维分布相匹配, 从而证明它存在最大误差的预测器( 据知梯度方法会趋向方向一致 ), 我们用它来论证,对于尽可能减少预期的偏差,, 对于迄今为止所研究的简单线性线性回归损失 。