High-dimensional statistical learning (HDSL) has wide applications in data analysis, operations research, and decision-making. Despite the availability of multiple theoretical frameworks, most existing HDSL schemes stipulate the following two conditions: (a) the sparsity, and (b) the restricted strong convexity (RSC). This paper generalizes both conditions via the use of the folded concave penalty (FCP). More specifically, we consider an M-estimation problem where (i) the (conventional) sparsity is relaxed into the approximate sparsity and (ii) the RSC is completely absent. We show that the FCP-based regularization leads to poly-logarithmic sample complexity; the training data size is only required to be poly-logarithmic in the problem dimensionality. This finding can facilitate the analysis of two important classes of models that are currently less understood: the high-dimensional nonsmooth learning and the (deep) neural networks (NN). For both problems, we show that the poly-logarithmic sample complexity can be maintained. In particular, our results indicate that the generalizability of NNs under over-parameterization can be theoretically ensured with the aid of regularization.
翻译:高层次统计学习(HDSL)在数据分析、业务研究和决策方面有着广泛的应用。尽管存在多种理论框架,但大多数现有的HDSL计划都规定了以下两个条件:(a) 宽度,和(b) 限制强共性(RSC) 。本文通过使用折叠共结罚款(FCP)概括了两种条件。更具体地说,我们认为,在(i) (常规)宽度放松到大致的广度和(ii) RSC完全不存在的情况下,存在着一个M-估计问题。我们表明,基于FCP的正规化导致多对数抽样复杂性;培训数据的规模仅要求为问题维度的多元对数。这一发现有助于分析目前不太理解的两大类重要模型:高维非光学和(深度)神经网络(NNN)。关于这两个问题,我们表明,基于FCP的正规化的样本复杂性可以维持多对数度的样本复杂性。特别是,我们的结果表明,在质论下,对NM的正规化在理论上的可靠性可以保证。