Gaussian processes (GPs) are powerful but computationally expensive machine learning models, requiring an estimate of the kernel covariance matrix for every prediction. In large and complex domains, such as graphs, sets, or images, the choice of suitable kernel can also be non-trivial to determine, providing an additional obstacle to the learning task. Over the last decade, these challenges have resulted in significant advances being made in terms of scalability and expressivity, exemplified by, e.g., the use of inducing points and neural network kernel approximations. In this paper, we propose inducing Gaussian process networks (IGN), a simple framework for simultaneously learning the feature space as well as the inducing points. The inducing points, in particular, are learned directly in the feature space, enabling a seamless representation of complex structured domains while also facilitating scalable gradient-based learning methods. We consider both regression and (binary) classification tasks and report on experimental results for real-world data sets showing that IGNs provide significant advances over state-of-the-art methods. We also demonstrate how IGNs can be used to effectively model complex domains using neural network architectures.
翻译:高斯进程(GPs)是强大但计算成本高昂的机器学习模型(GPs),需要估算每场预测的内核共变矩阵。在图、集或图像等大型和复杂领域,合适的内核的选择也可能不是三重决定,为学习任务提供了额外的障碍。在过去十年中,这些挑战导致在可缩放性和表达性方面取得重大进展,例如,使用导出点和神经网络内核近距离等,本文建议引导高斯进程网络(Gausian process network),这是同时学习地貌空间和导出点的简单框架。特别是,导出点在地貌空间直接学习,使复杂结构化领域能够无缝地代表,同时促进可缩放的梯度学习方法。我们既考虑回归性和(二进制)分类任务,又报告真实世界数据集的实验结果,表明IGNs提供了超越状态方法的重大进展。我们还演示了如何使用IGNs有效地模拟复杂的网络结构。