Traditional Bayesian approaches for model uncertainty quantification rely on notoriously difficult processes of marginalization over each network parameter to estimate its probability density function (PDF). Our hypothesis is that internal layer outputs of a trained neural network contain all of the information related to both its mapping function (quantified by its weights) as well as the input data distribution. We therefore propose a framework for predictive uncertainty quantification of a trained neural network that explicitly estimates the PDF of its raw prediction space (before activation), p(y'|x,w), which we refer to as the model PDF, in a Gaussian reproducing kernel Hilbert space (RKHS). The Gaussian RKHS provides a localized density estimate of p(y'|x,w), which further enables us to utilize gradient based formulations of quantum physics to decompose the model PDF in terms of multiple local uncertainty moments that provide much greater resolution of the PDF than the central moments characterized by Bayesian methods. This provides the framework with a better ability to detect distributional shifts in test data away from the training data PDF learned by the model. We evaluate the framework against existing uncertainty quantification methods on benchmark datasets that have been corrupted using common perturbation techniques. The kernel framework is observed to provide model uncertainty estimates with much greater precision based on the ability to detect model prediction errors.
翻译:我们的假设是,经过训练的神经网络的内部层输出包含与其绘图功能(按重量量化)以及输入数据分布有关的所有信息。因此,我们提议了一个框架,用于预测一个经过训练的神经网络的不确定性量化,该网络明确估计其原始预测空间(在激活前)的PDF, p(y' ⁇ x,w),我们称之为模型PDF,用于估算其概率密度功能(RKHS)的模型PDF。Gausian RKHS提供了对p(y' ⁇ x,w)的局部密度估计,这进一步使我们能够利用量子物理的梯度公式,将模型PDFF分解为多个局部不确定时刻,这些时候对PDF的分辨率比Bayesian方法所描述的中心时刻要高得多。这为框架提供了一种更好的能力,以检测测试数据从模型所学的培训数据PDFF的分布变化。我们用通用的精确度估算法评估了比通用的精确度框架更精确性。我们用观察到的精确度来评估现有不确定性的精确度框架提供了更多的基准。