Gradient Boosting Machines (GBMs) have demonstrated remarkable success in solving diverse problems by utilizing Taylor expansions in functional space. However, achieving a balance between performance and generality has posed a challenge for GBMs. In particular, gradient descent-based GBMs employ the first-order Taylor expansion to ensure applicability to all loss functions, while Newton's method-based GBMs use positive Hessian information to achieve superior performance at the expense of generality. To address this issue, this study proposes a new generic Gradient Boosting Machine called Trust-region Boosting (TRBoost). In each iteration, TRBoost uses a constrained quadratic model to approximate the objective and applies the Trust-region algorithm to solve it and obtain a new learner. Unlike Newton's method-based GBMs, TRBoost does not require the Hessian to be positive definite, thereby allowing it to be applied to arbitrary loss functions while still maintaining competitive performance similar to second-order algorithms. The convergence analysis and numerical experiments conducted in this study confirm that TRBoost is as general as first-order GBMs and yields competitive results compared to second-order GBMs. Overall, TRBoost is a promising approach that balances performance and generality, making it a valuable addition to the toolkit of machine learning practitioners.
翻译:梯度提升机(GBM)通过在函数空间中利用泰勒展开,在解决各种问题方面取得了显著的成功。然而,在性能和通用性之间取得平衡一直是GBM面临的挑战。特别是,基于梯度下降的GBM采用一阶泰勒展开来确保适用于所有损失函数,而基于牛顿法的GBM使用正定的海森矩阵信息以在性能上取得优异结果,但通用性(generalization)并不好。为了解决这个问题,本研究提出了一种新的通用梯度提升机——Trust-region Boosting(TRBoost)。在每次迭代中,TRBoost使用受限二次模型来逼近目标,并应用信任区域算法求解它以获得一个新的学习器。与基于牛顿法的GBM不同,TRBoost不要求海森矩阵是正定的,从而使其可以应用于任意损失函数,同时仍然保持类似于二阶算法的竞争性能。本研究进行的收敛性分析和数值实验证实,TRBoost和一阶GBM一样通用,并与二阶GBM相比具有竞争力。总的来说,TRBoost是一种平衡性能和通用性的有前途的方法,使其成为机器学习从业者工具箱中有价值的补充。