In this paper, we study the problem of early stopping for iterative learning algorithms in a reproducing kernel Hilbert space (RKHS) in the nonparametric regression framework. In particular, we work with the gradient descent and (iterative) kernel ridge regression algorithms. We present a data-driven rule to perform early stopping without a validation set that is based on the so-called minimum discrepancy principle. This method enjoys only one assumption on the regression function: it belongs to a reproducing kernel Hilbert space (RKHS). The proposed rule is proved to be minimax-optimal over different types of kernel spaces, including finite-rank and Sobolev smoothness classes. The proof is derived from the fixed-point analysis of the localized Rademacher complexities, which is a standard technique for obtaining optimal rates in the nonparametric regression literature. In addition to that, we present simulation results on artificial datasets that show the comparable performance of the designed rule with respect to other stopping rules such as the one determined by V-fold cross-validation.
翻译:在本文中,我们研究了在非参数回归框架内复制核心的Hilbert空间(RKHS)中早期停止进行迭代学习算法的问题,特别是,我们与梯度下移和(临时)内核回归算法合作,我们提出了一个数据驱动规则,以便在没有基于所谓最小差异原则的验证集的情况下进行早期停止,这一方法在回归函数上只有一个假设:它属于复制核心Hilbert空间(RKHS),拟议规则被证明是对不同类型内核空间,包括定级和索博勒夫平滑等级的小型最大最佳规则。证据来自局部Rademacher复杂性的定点分析,这是在非参数回归文献中获得最佳率的标准方法。此外,我们还介绍了人工数据集的模拟结果,该模型显示设计规则与V倍交叉校准确定的其他停止规则具有可比性。