We introduce a simple and scalable method for training Gaussian process (GP) models that exploits cross-validation and nearest neighbor truncation. To accommodate binary and multi-class classification we leverage P\`olya-Gamma auxiliary variables and variational inference. In an extensive empirical comparison with a number of alternative methods for scalable GP regression and classification, we find that our method offers fast training and excellent predictive performance. We argue that the good predictive performance can be traced to the non-parametric nature of the resulting predictive distributions as well as to the cross-validation loss, which provides robustness against model mis-specification.
翻译:我们采用一种简单、可扩展的方法来培训高斯进程模型,利用交叉验证和最近的邻里脱轨。为了适应二进制和多级分类,我们利用了P ⁇ olya-Gamma的辅助变量和可变推论。在与可伸缩的Gaussian回归和分类的若干替代方法进行的广泛经验比较中,我们发现我们的方法提供了快速培训和出色的预测性能。我们争辩说,良好的预测性能可以追溯到由此产生的预测性分布的非参数性质以及交叉验证性损失,后者提供了抵御模型误差的稳健性。