Gaussian processes (GPs) are non-parametric regression engines with a long history. They are often overlooked in modern machine learning contexts because of scalability issues: regression for traditional GP kernels are $\mathcal{O}(N^3)$ where $N$ is the size of the dataset. One of a number of scalable GP approaches is the Karhunen-Lo\'eve (KL) decomposed kernel BSS-ANOVA, developed in 2009. It is $\mathcal{O}(NP)$ in training and $\mathcal{O}(P)$ per point in prediction, where $P$ is the number of terms in the ANOVA / KL expansion. A new method of forward variable selection, quickly and effectively limits the number of terms, yielding a method with competitive accuracies, training and inference times for large tabular datasets. The new algorithm balances model fidelity with model complexity using Bayesian and Akaike information criteria (BIC/AIC). The inference speed and accuracy makes the method especially useful for modeling dynamic systems in a model-free manner, by modeling the derivative in a dynamic system as a static problem, then integrating the learned dynamics using a high-order scheme. The methods are demonstrated on a `Susceptible, Infected, Recovered' (SIR) toy problem, with the transmissibility used as forcing function, along with the `Cascaded Tanks' benchmark dataset. Comparisons on the static prediction of derivatives are made with a Random Forest and Residual Neural Network, while for the timeseries prediction comparisons are made with LSTM and GRU recurrent neural networks. The GP outperforms the other methods in all modeling tasks on accuracy, while (in the case of the neural networks) performing many orders of magnitude fewer calculations. For the SIR test, which involved prediction for a set of forcing functions qualitatively different from those appearing in the training set, the GP captured the correct dynamics while the neural networks failed to do so.
翻译:高斯进程( GPs) 是具有悠久历史的非参数回归引擎 。 在现代机器学习环境中, 它们经常被忽略, 因为可缩缩缩问题: 传统的 GP 内核的退缩是$\ mathca{O}( N3) 美元, 美元是数据集的大小。 一系列可缩放的GP 方法之一是 Karhunen- Lo\'eve ( KL), 由2009年开发的 Karhunen- Lo\'eve (KLL) 内核流 BSS- ANOVA (NP), 在培训中和 $\ mathcals (PNP) 中被忽略。 在预测中, 传统的 GGNOVA/ KL 扩展中, $P$ 的退缩是 。 一种新的变量选择方法, 迅速有效地限制语系, 产生一种具有竞争性缩略释法的方法,, 用于大型的 模型的解算法 。 在模型中, 解算法中, 新算法中, 的 Ration 调调调调的调调调调调调调调调调调调调调调调调调调调调调调调调调调调调和调和调和调调调调调调调调调调调调调调调调调调调调调调调调调调调调调调调调调调调 。