Gradient boosting is a prediction method that iteratively combines weak learners to produce a complex and accurate model. From an optimization point of view, the learning procedure of gradient boosting mimics a gradient descent on a functional variable. This paper proposes to build upon the proximal point algorithm, when the empirical risk to minimize is not differentiable, in order to introduce a novel boosting approach, called proximal boosting. It comes with a companion algorithm inspired by [1] and called residual proximal boosting, which is aimed at better controlling the approximation error. Theoretical convergence is proved for these two procedures under different hypotheses on the empirical risk and advantages of leveraging proximal methods for boosting are illustrated by numerical experiments on simulated and real-world data. In particular, we exhibit a favorable comparison over gradient boosting regarding convergence rate and prediction accuracy.
翻译:梯度递增是一种预测方法,迭代地将弱学习者结合成一个复杂而准确的模型。 从优化的角度看,梯度递增的学习程序模仿功能变量上的梯度递减。本文件提议在模拟和现实世界数据的数字实验中,以近似点算法为基础,在模拟和现实世界数据中,模拟利用准加速方法的经验风险和优势证明了这两种程序理论趋同。特别是,我们对趋同率和预测准确性方面的梯度推增进行了有利的比较。