We consider the problem of computing a Gaussian approximation to the posterior distribution of a parameter given a large number N of observations and a Gaussian prior, when the dimension of the parameter d is also large. To address this problem we build on a recently introduced recursive algorithm for variational Gaussian approximation of the posterior, called recursive variational Gaussian approximation (RVGA), which is a single pass algorithm, free of parameter tuning. In this paper, we consider the case where the parameter dimension d is high, and we propose a novel version of RVGA that scales linearly in the dimension d (as well as in the number of observations N), and which only requires linear storage capacity in d. This is afforded by the use of a novel recursive expectation maximization (EM) algorithm applied for factor analysis introduced herein, to approximate at each step the covariance matrix of the Gaussian distribution conveying the uncertainty in the parameter. The approach is successfully illustrated on the problems of high dimensional least-squares and logistic regression, and generalized to a large class of nonlinear models.
翻译:我们考虑在参数的维数d也很大时,给定大量N个观测值和高斯先验的情况下,如何计算参数后验分布的高斯近似。为了解决这个问题,我们在一种最近引入的递归算法,即递归变分高斯近似(RVGA)的基础上建立,该算法是一个单通算法,无需参数调整。在本文中,我们考虑参数维数d很高的情况,并提出一种新颖的RVGA版本,其在维数d(以及观测值N个数)上的规模呈线性增长,并且仅需要线性的存储容量。这是通过用于因子分析的一种新型递归期望极大(EM)算法来近似在每一步中传达参数不确定性的高斯分布的协方差矩阵所实现的。该方法在高维最小二乘和逻辑回归问题上得到了成功应用,并推广到了一大类非线性模型。