We propose to compute a sparse approximate inverse Cholesky factor $L$ of a dense covariance matrix $\Theta$ by minimizing the Kullback-Leibler divergence between the Gaussian distributions $\mathcal{N}(0, \Theta)$ and $\mathcal{N}(0, L^{-\top} L^{-1})$, subject to a sparsity constraint. Surprisingly, this problem has a closed-form solution that can be computed efficiently, recovering the popular Vecchia approximation in spatial statistics. Based on recent results on the approximate sparsity of inverse Cholesky factors of $\Theta$ obtained from pairwise evaluation of Green's functions of elliptic boundary-value problems at points $\{x_{i}\}_{1 \leq i \leq N} \subset \mathbb{R}^{d}$, we propose an elimination ordering and sparsity pattern that allows us to compute $\epsilon$-approximate inverse Cholesky factors of such $\Theta$ in computational complexity $\mathcal{O}(N \log(N/\epsilon)^d)$ in space and $\mathcal{O}(N \log(N/\epsilon)^{2d})$ in time. To the best of our knowledge, this is the best asymptotic complexity for this class of problems. Furthermore, our method is embarrassingly parallel, automatically exploits low-dimensional structure in the data, and can perform Gaussian-process regression in linear (in $N$) space complexity. Motivated by the optimality properties of our methods, we propose methods for applying it to the joint covariance of training and prediction points in Gaussian-process regression, greatly improving stability and computational cost. Finally, we show how to apply our method to the important setting of Gaussian processes with additive noise, sacrificing neither accuracy nor computational complexity.
翻译:我们提议在不显眼的限制下, 计算一个微小的Choolesky 低度的低度变异因子 $\ Theta$, 以最小化的基质变异基基 $\ Theta$( 0, \ theta) $\ mathcal{N} 美元和 $\ mathcal{N} 美元之间的差价。 令人惊讶的是, 这个问题有一个封闭式的解决方案, 可以有效计算, 恢复广受欢迎的Vecchia 快速变异基基 $\ Theta$。 根据最近的结果, 粗度变异性变异法的基数的基数 $\ the rodeal democredical ral resmation $_ $N>_ the mocideal_ commology} 将绿色的边界值的功能评估结果 $x% 1\ \ \ \ ileq}\ asubs mathbretial res res mission to rodustration.