Boosting is one of the most significant developments in machine learning. This paper studies the rate of convergence of $L_2$Boosting, which is tailored for regression, in a high-dimensional setting. Moreover, we introduce so-called \textquotedblleft post-Boosting\textquotedblright. This is a post-selection estimator which applies ordinary least squares to the variables selected in the first stage by $L_2$Boosting. Another variant is \textquotedblleft Orthogonal Boosting\textquotedblright\ where after each step an orthogonal projection is conducted. We show that both post-$L_2$Boosting and the orthogonal boosting achieve the same rate of convergence as LASSO in a sparse, high-dimensional setting. We show that the rate of convergence of the classical $L_2$Boosting depends on the design matrix described by a sparse eigenvalue constant. To show the latter results, we derive new approximation results for the pure greedy algorithm, based on analyzing the revisiting behavior of $L_2$Boosting. We also introduce feasible rules for early stopping, which can be easily implemented and used in applied work. Our results also allow a direct comparison between LASSO and boosting which has been missing from the literature. Finally, we present simulation studies and applications to illustrate the relevance of our theoretical results and to provide insights into the practical aspects of boosting. In these simulation studies, post-$L_2$Boosting clearly outperforms LASSO.
翻译:推动是机器学习中最重要的进展之一 。 本文在高维环境下研究了用于回归的 $L_ 2$ Boosting 的趋同率。 此外, 我们引入了所谓的“ textquotledblleft post- Boosting” 和“ textgreedblight” 。 这是一个选后估算器, 将普通最低方块应用到第一个阶段由 $L_ 2$ Boosting 选择的变量上。 另一个变量是\ textblentblordleft Orthogotal lobsting\ leftalblight\, 每一步后都会进行一个或多层次的预测。 我们显示的是, 后一步骤的L_ 2$ Boostem 和 othobalpal 的提振速度都达到了与 LASSSO相同的趋同率。 我们显示经典的趋同率速度取决于设计矩阵, 由稀少的值常数常数描述。 为了显示后一结果, 我们从分析的纯贪算算结果中得出新的结果, 我们的L_ 2号的重新分析的推算法, 也可以在使用这些研究中, 直接推算结果中, 开始使用这些结果, 我们的L_ Boal_ 和最终应用的推算结果, 和我们使用这些推算法可以让这些结果。