Deterministic neural networks (NNs) are increasingly being deployed in safety critical domains, where calibrated, robust, and efficient measures of uncertainty are crucial. In this paper, we propose a novel method for training non-Bayesian NNs to estimate a continuous target as well as its associated evidence in order to learn both aleatoric and epistemic uncertainty. We accomplish this by placing evidential priors over the original Gaussian likelihood function and training the NN to infer the hyperparameters of the evidential distribution. We additionally impose priors during training such that the model is regularized when its predicted evidence is not aligned with the correct output. Our method does not rely on sampling during inference or on out-of-distribution (OOD) examples for training, thus enabling efficient and scalable uncertainty learning. We demonstrate learning well-calibrated measures of uncertainty on various benchmarks, scaling to complex computer vision tasks, as well as robustness to adversarial and OOD test samples.
翻译:确定性神经网络(NNs)正越来越多地部署在安全关键领域,在这个关键领域,校准、稳健和高效的不确定性措施至关重要。在本文件中,我们提出一种新的方法,用于培训非巴伊西亚NNS,以估计连续目标及其相关证据,从而既学习感官不确定性,也学习认知性不确定性。我们通过在最初高斯概率功能上设定证据前科,培训NN,以推断证据分布的超参数。我们还在培训期间规定了前科,以便在预测证据与正确产出不一致时,使模型正规化。我们的方法并不依赖在推断或分配外实例中进行取样,从而使得能够高效和可扩展的不确定性学习。我们展示了在各种基准上对各种不确定性的衡量方法,将尺度缩放到复杂的计算机视觉任务上,并展示了对立和 OOODD测试样品的坚固性。