Mixtures of experts have become an indispensable tool for flexible modelling in a supervised learning context, and sparse Gaussian processes (GP) have shown promise as a leading candidate for the experts in such models. In this article, we propose to design the gating network for selecting the experts from such mixtures of sparse GPs using a deep neural network (DNN). Furthermore, a fast one pass algorithm called Cluster-Classify-Regress (CCR) is leveraged to approximate the maximum a posteriori (MAP) estimator extremely quickly. This powerful combination of model and algorithm together delivers a novel method which is flexible, robust, and extremely efficient. In particular, the method is able to outperform competing methods in terms of accuracy and uncertainty quantification. The cost is competitive on low-dimensional and small data sets, but is significantly lower for higher-dimensional and big data sets. Iteratively maximizing the distribution of experts given allocations and allocations given experts does not provide significant improvement, which indicates that the algorithm achieves a good approximation to the local MAP estimator very fast. This insight can be useful also in the context of other mixture of experts models.
翻译:专家混合已成为在有监督的学习环境中进行灵活建模的不可或缺的工具,而稀有的高斯进程(GP)作为这类模型专家的主要候选人表现出了前景。在本条中,我们提议设计一个格子网络,利用深神经网络(DNN)从这种稀多的GP混合物中挑选专家。此外,一个称为Croup-Clasti-Regrest(CCCR)的快速传算法被利用来极快地接近事后估测员(MAP)的最大比例。这种强大的模型和算法组合提供了一种灵活、有力和极有效率的新方法。特别是,该方法能够在准确性和不确定性的量化方面超越相互竞争的方法。对于低维和小型数据集来说,成本是竞争性的,但对于高维和大数据集来说则要低得多。 尽量扩大专家的分配和分配并不能提供显著的改进,这表明该算法能够非常迅速地接近当地MAP估测算师。这种洞察力在其他模型中也很有用。