This work considers variational Bayesian inference as an inexpensive and scalable alternative to a fully Bayesian approach in the context of sparsity-promoting priors. In particular, the priors considered arise from scale mixtures of Normal distributions with a generalized inverse Gaussian mixing distribution. This includes the variational Bayesian LASSO as an inexpensive and scalable alternative to the Bayesian LASSO introduced in [65]. It also includes a family of priors which more strongly promote sparsity. For linear models the method requires only the iterative solution of deterministic least squares problems. Furthermore, for p unknown covariates the method can be implemented exactly online with a cost of $O(p^3)$ in computation and $O(p^2)$ in memory per iteration -- in other words, the cost per iteration is independent of n, and in principle infinite data can be considered. For large $p$ an approximation is able to achieve promising results for a cost of $O(p)$ per iteration, in both computation and memory. Strategies for hyper-parameter tuning are also considered. The method is implemented for real and simulated data. It is shown that the performance in terms of variable selection and uncertainty quantification of the variational Bayesian LASSO can be comparable to the Bayesian LASSO for problems which are tractable with that method, and for a fraction of the cost. The present method comfortably handles $n = 65536$, $p = 131073$ on a laptop in less than 30 minutes, and $n = 10^5$, $p = 2.1 \times 10^6$ overnight.
翻译:这项工作认为,变式贝雅的推算是完全巴伊西亚方法的一种廉价和可缩放的替代方法。 特别是, 所考虑的先验是正常分配的比重混合物, 普遍反戈斯混合分布, 包括变式巴伊西亚LASSO, 是贝伊西亚LASO的廉价和可缩放的替代方法, 在 [65] 中采用。 它还包括一个更能促进恐慌的先验家庭。 对于直线模型来说,该方法只要求确定性最低平方问题的迭代解决方案。 此外, 未知的共变式方法可以完全在线实施, 计算中成本为O( p3)$, 记忆中成本为O( p2美元), 记忆中成本为O( p2美元)。 换句话说, 每升价成本是n,原则上可以考虑。 如果高一美元, 近于13美元, 每平价的处理成本为美元, 在计算和记忆中,每升一美元。 超常价的SLA值 5 的计算和可变性SOSO值方法, 在模拟中, 成本和变数方法中显示, 。 可变性SOSO= 。 可变数的方法是 。 。