Despite their compelling theoretical properties, Bayesian neural networks (BNNs) tend to perform worse than frequentist methods in classification-based uncertainty quantification (UQ) tasks such as out-of-distribution (OOD) detection and dataset-shift robustness. In this work, based on empirical findings in prior works, we hypothesize that this issue is due to the avoidance of Bayesian methods in the so-called "OOD training" -- a family of techniques for incorporating OOD data during training process, which has since been an integral part of state-of-the-art frequentist UQ methods. To validate this, we treat OOD data as a first-class citizen in BNN training by exploring four different ways of incorporating OOD data in Bayesian inference. We show in extensive experiments that OOD-trained BNNs are competitive to, if not better than recent frequentist baselines. This work thus provides strong baselines for future work in both Bayesian and frequentist UQ.
翻译:尽管有令人信服的理论特性,但巴耶斯神经网络(BNNs)在基于分类的不确定性量化(UQ)任务方面往往比经常采用的方法差,这些任务包括分解(OOD)检测和数据集易变的稳健性。在这项工作中,根据以往工作的经验调查结果,我们假设,这一问题是由于在所谓的“OOOD培训”中避免了巴耶斯方法造成的。 “OOOD培训”是培训过程中将OOD数据纳入OD数据的一套技术,自此成为最先进常者UQ方法的一个组成部分。为了证实这一点,我们在BNN培训中将OOD数据作为一流公民对待,探索将OOD数据纳入巴耶斯误判的四种不同方法。我们在广泛的实验中显示,OD训练的BNNs具有竞争力,即使不比最近的经常者基线更好。因此,这项工作为巴耶斯和经常者UQ的未来工作提供了强有力的基准。