Choosing a proper set of kernel functions is an important problem in learning Gaussian Process (GP) models since each kernel structure has different model complexity and data fitness. Recently, automatic kernel composition methods provide not only accurate prediction but also attractive interpretability through search-based methods. However, existing methods suffer from slow kernel composition learning. To tackle large-scaled data, we propose a new sparse approximate posterior for GPs, MultiSVGP, constructed from groups of inducing points associated with individual additive kernels in compositional kernels. We demonstrate that this approximation provides a better fit to learn compositional kernels given empirical observations. We also theoretically justification on error bound when compared to the traditional sparse GP. In contrast to the search-based approach, we present a novel probabilistic algorithm to learn a kernel composition by handling the sparsity in the kernel selection with Horseshoe prior. We demonstrate that our model can capture characteristics of time series with significant reductions in computational time and have competitive regression performance on real-world data sets.
翻译:选择一套适当的内核函数是学习高斯进程模型中的一个重要问题,因为每个内核结构的模型复杂程度和数据适合性不同。最近,自动内核组成方法不仅提供了准确的预测,而且通过搜索方法提供了有吸引力的解释性。然而,现有方法受到慢内核成分学习的影响。为了解决大规模数据问题,我们建议为GP提出一个新的稀疏的近似后部,MultiSVGP, 由组合内核中与单个添加内核相关的导点组组成。我们证明,这种近似比较适合根据实证观察来学习组成内核。我们还从理论上解释了与传统的稀有的GP相比存在哪些错误。与基于搜索的方法不同,我们提出了一个新的概率算法,通过处理前马座的内核选择中的孔来学习内核成分。我们证明,我们的模型可以捕捉时间序列特征,计算时间大幅减少,并在现实世界数据集上具有竞争性的回归性表现。