Understanding generalization and estimation error of estimators for simple models such as linear and generalized linear models has attracted a lot of attention recently. This is in part due to an interesting observation made in machine learning community that highly over-parameterized neural networks achieve zero training error, and yet they are able to generalize well over the test samples. This phenomenon is captured by the so called double descent curve, where the generalization error starts decreasing again after the interpolation threshold. A series of recent works tried to explain such phenomenon for simple models. In this work, we analyze the asymptotics of estimation error in ridge estimators for convolutional linear models. These convolutional inverse problems, also known as deconvolution, naturally arise in different fields such as seismology, imaging, and acoustics among others. Our results hold for a large class of input distributions that include i.i.d. features as a special case. We derive exact formulae for estimation error of ridge estimators that hold in a certain high-dimensional regime. We show the double descent phenomenon in our experiments for convolutional models and show that our theoretical results match the experiments.
翻译:对线性模型和一般线性模型等简单模型的估测者一般化和估计错误的理解和估计错误最近引起了许多注意。这在一定程度上是由于机器学习界的一项有意思的观察,即高度偏差的神经神经网络没有受过任何训练,然而它们却能够对测试样品进行广泛概括。这种现象被所谓的双向曲线所捕捉,在这种曲线中,一般化错误在内推临界值之后又开始减少。最近的一系列工作试图为简单模型解释这种现象。在这项工作中,我们分析了革命性线性模型山脊估计错误的暂时性。这些反向问题,也被称为演化问题,自然出现在诸如地震学、成像和声学等不同领域。我们的结果支持了包括i.i.d.特征在内的一大批输入分布,作为一个特殊案例。我们为估计某一高度系统中的脊心估计器的误差得出了精确的公式。我们展示了我们革命性模型实验中的双重血统现象,并展示了我们的理论结果匹配实验结果。