Overparametrized neural networks tend to perfectly fit noisy training data yet generalize well on test data. Inspired by this empirical observation, recent work has sought to understand this phenomenon of benign overfitting or harmless interpolation in the much simpler linear model. Previous theoretical work critically assumes that either the data features are statistically independent or the input data is high-dimensional; this precludes general nonparametric settings with structured feature maps. In this paper, we present a general and flexible framework for upper bounding regression and classification risk in a reproducing kernel Hilbert space. A key contribution is that our framework describes precise sufficient conditions on the data Gram matrix under which harmless interpolation occurs. Our results recover prior independent-features results (with a much simpler analysis), but they furthermore show that harmless interpolation can occur in more general settings such as features that are a bounded orthonormal system. Furthermore, our results show an asymptotic separation between classification and regression performance in a manner that was previously only shown for Gaussian features.
翻译:过度平衡的神经网络往往完全适合杂乱的培训数据,但是却对测试数据进行了全面概括。根据这一经验观察,最近的工作力求理解在简单得多的线性模型中良性地超配或无害的内插现象。先前的理论工作严格地假定数据特征在统计上是独立的,或输入数据是高维的;这排除了带有结构化特征图的一般非对称设置。在本文中,我们提出了一个在再生产核心内尔·希尔伯特空间上层约束回归和分类风险的一般和灵活的框架。一个关键贡献是,我们的框架描述了数据格拉姆矩阵的准确条件,在这个矩阵中,无害的内插。我们的结果恢复了先前的独立特征结果(分析简单得多 ), 但是它们进一步表明无害的内插可能发生在更一般的环境下,例如连接的异常系统特征。此外,我们的结果显示分类和回归性表现之间有不严谨的分离,其方式以前只为戈斯特征所显示。