Deep learning methods operate in regimes that defy the traditional statistical mindset. Neural network architectures often contain more parameters than training samples, and are so rich that they can interpolate the observed labels, even if the latter are replaced by pure noise. Despite their huge complexity, the same architectures achieve small generalization error on real data. This phenomenon has been rationalized in terms of a so-called `double descent' curve. As the model complexity increases, the test error follows the usual U-shaped curve at the beginning, first decreasing and then peaking around the interpolation threshold (when the model achieves vanishing training error). However, it descends again as model complexity exceeds this threshold. The global minimum of the test error is found above the interpolation threshold, often in the extreme overparametrization regime in which the number of parameters is much larger than the number of samples. Far from being a peculiar property of deep neural networks, elements of this behavior have been demonstrated in much simpler settings, including linear regression with random covariates. In this paper we consider the problem of learning an unknown function over the $d$-dimensional sphere $\mathbb S^{d-1}$, from $n$ i.i.d. samples $(\boldsymbol x_i, y_i)\in \mathbb S^{d-1} \times \mathbb R$, $i\le n$. We perform ridge regression on $N$ random features of the form $\sigma(\boldsymbol w_a^{\mathsf T} \boldsymbol x)$, $a\le N$. This can be equivalently described as a two-layers neural network with random first-layer weights. We compute the precise asymptotics of the test error, in the limit $N,n,d\to \infty$ with $N/d$ and $n/d$ fixed. This provides the first analytically tractable model that captures all the features of the double descent phenomenon without assuming ad hoc misspecification structures.
翻译:深度学习方法在无视传统统计思维的制度中运作。 神经网络架构通常包含比培训样本更多的参数, 并且非常丰富, 以至于可以对观察到的标签进行内推, 即使后者被纯噪音所取代。 尽管这些结构非常复杂, 却在真实数据中实现了小的概括错误。 这种现象在所谓的“ 双向” 曲线上得到了合理化。 随着模型复杂性的增加, 测试错误遵循了通常的 U 形状曲线, 首先下降, 然后在内推门槛周围达到峰值( 当模型实现了培训错误消失时 ) 。 然而, 当模型复杂度超过这个阈值时, 它又会再次出现内插。 全球测试错误的最低值超过了内插阈值的阈值, 其中参数的数量远大于样本的数量。 远不是深层神经网络的奇特特性, 而这种行为的元素可以在更简单的环境中表现出来, 包括随机的直线回归( 当模型完成后, 以美元- 美元- 美元- 美元- 美元- 平面的 测试 。