Bayesian formulations of deep learning have been shown to have compelling theoretical properties and offer practical functional benefits, such as improved predictive uncertainty quantification and model selection. The Laplace approximation (LA) is a classic, and arguably the simplest family of approximations for the intractable posteriors of deep neural networks. Yet, despite its simplicity, the LA is not as popular as alternatives like variational Bayes or deep ensembles. This may be due to assumptions that the LA is expensive due to the involved Hessian computation, that it is difficult to implement, or that it yields inferior results. In this work we show that these are misconceptions: we (i) review the range of variants of the LA including versions with minimal cost overhead; (ii) introduce "laplace", an easy-to-use software library for PyTorch offering user-friendly access to all major flavors of the LA; and (iii) demonstrate through extensive experiments that the LA is competitive with more popular alternatives in terms of performance, while excelling in terms of computational cost. We hope that this work will serve as a catalyst to a wider adoption of the LA in practical deep learning, including in domains where Bayesian approaches are not typically considered at the moment.
翻译:深入学习的贝叶斯人配方已被证明具有令人信服的理论特性,并提供了实际功能效益,例如改进了预测不确定性的量化和模型选择。拉普尔近似(LA)是一个经典的,可以说是深神经网络中最简单的近似组合。然而,尽管简便,拉普尔的配方不像变异贝耶斯或深层合成物等替代物那么受欢迎。这可能是由于假设LA由于涉及的赫森计算方法而昂贵,难以执行,或产生低效的结果。在这项工作中,我们希望这些是错误的:我们(一)审查LA的各种变方,包括成本最低的版本;(二)为PyTorrch引入“place”,一个方便使用的软件图书馆,提供方便用户的LA所有主要口味;以及(三)通过广泛的实验,证明LA在业绩方面与更受欢迎的替代物竞争,同时在计算成本方面优异。我们希望这项工作能成为在实际深层次的学习领域,包括通常的Bay地区,作为更广泛地采用LA的催化剂。