The Bayesian additive regression trees (BART) model is an ensemble method extensively and successfully used in regression tasks due to its consistently strong predictive performance and its ability to quantify uncertainty. BART combines "weak" tree models through a set of shrinkage priors, whereby each tree explains a small portion of the variability in the data. However, the lack of smoothness and the absence of a covariance structure over the observations in standard BART can yield poor performance in cases where such assumptions would be necessary. We propose Gaussian processes Bayesian additive regression trees (GP-BART) as an extension of BART which assumes Gaussian process (GP) priors for the predictions of each terminal node among all trees. We illustrate our model on simulated and real data and compare its performance to traditional modelling approaches, outperforming them in many scenarios. An implementation of our method is available in the R package rGPBART available at: https://github.com/MateusMaiaDS/gpbart
翻译:Bayesian 添加性回归树(BART)模型是广泛和成功地用于回归任务的一种混合方法,因为它具有一贯强大的预测性能和量化不确定性的能力。BART将“弱”树模型通过一套缩水前期结合,每棵树解释数据变异的一小部分。然而,标准BART中观测结果缺乏顺畅和共变结构,在需要这种假设的情况下可能产生不良的性能。我们建议Gaussian进程Bayesian 添加性回归树(GP-BART)作为BART的延伸,它假定Gaussian进程(GP)之前预测所有树木的每个终点节点。我们用模拟和真实数据模型来说明我们的模型,并将其性能与传统的建模方法相比较,在许多情景中优于这些模型方法。我们的方法的实施见RGPBART,网址:https://github.com/Mateus MaiaDS/gpbart。 https://gthub.com/Mateus MaiaDS/gpbart。