Gradient boosting machines (GBMs) based on decision trees consistently demonstrate state-of-the-art results on regression and classification tasks with tabular data, often outperforming deep neural networks. However, these models do not provide well-calibrated predictive uncertainties, which prevents their use for decision making in high-risk applications. The Bayesian treatment is known to improve predictive uncertainty calibration, but previously proposed Bayesian GBM methods are either computationally expensive, or resort to crude approximations. Variational inference is often used to implement Bayesian neural networks, but is difficult to apply to GBMs, because the decision trees used as weak learners are non-differentiable. In this paper, we propose to implement Bayesian GBMs using variational inference with soft decision trees, a fully differentiable alternative to standard decision trees introduced by Irsoy et al. Our experiments demonstrate that variational soft trees and variational soft GBMs provide useful uncertainty estimates, while retaining good predictive performance. The proposed models show higher test likelihoods when compared to the state-of-the-art Bayesian GBMs in 7/10 tabular regression datasets and improved out-of-distribution detection in 5/10 datasets.
翻译:基于决策树的梯度加速器(GBM)以决策树为基础,一贯以表单数据显示回归和分类任务的最新最新结果,通常优于深神经网络,但这些模型并不提供精确的预测不确定性,无法在高风险应用中用于决策。贝叶斯治疗已知可以改进预测不确定性的校准,但先前提出的巴伊西亚GBM方法要么计算成本昂贵,要么采用粗略的近似法。变化式推论常常用于实施贝伊斯神经网络,但难以适用于GBMs,因为作为薄弱学习者使用的决策树是不可区分的。在本文件中,我们提议采用对软决策树的变异推论来应用Bayesian GBMs,这是Irsoy等人引入的完全不同的标准决策树。我们的实验表明,变软树和变软性软GBMs提供了有用的不确定性估计,同时保留了良好的预测性性能。在7/10年的Bayes Regals regress 中,拟议模型显示与州-art-Arformissionalations 5/10的GBMDregress 中改进的数据比较高的测试可能性更高。