The Delta method is a classical procedure for quantifying epistemic uncertainty in statistical models, but its direct application to deep neural networks is prevented by the large number of parameters $P$. We propose a low cost variant of the Delta method applicable to $L_2$-regularized deep neural networks based on the top $K$ eigenpairs of the Fisher information matrix. We address efficient computation of full-rank approximate eigendecompositions in terms of either the exact inverse Hessian, the inverse outer-products of gradients approximation or the so-called Sandwich estimator. Moreover, we provide a bound on the approximation error for the uncertainty of the predictive class probabilities. We observe that when the smallest eigenvalue of the Fisher information matrix is near the $L_2$-regularization rate, the approximation error is close to zero even when $K\ll P$. A demonstration of the methodology is presented using a TensorFlow implementation, and we show that meaningful rankings of images based on predictive uncertainty can be obtained for two LeNet-based neural networks using the MNIST and CIFAR-10 datasets. Further, we observe that false positives have on average a higher predictive epistemic uncertainty than true positives. This suggests that there is supplementing information in the uncertainty measure not captured by the classification alone.
翻译:三角洲方法是一种典型的程序,用于量化统计模型中的隐性不确定性,但是它直接应用于深神经网络却受到大量参数的阻碍。我们提议了一个适用于Fisher信息矩阵中以最高价值K$egenpair为基础的、以Fisher信息矩阵中最高价值K$2美元为基础的、正规化的深神经网络的三角洲方法的低成本变方。我们处理的是以赫森正好相反的赫森、梯度近似的反外产品或所谓的桑威奇估计器的反外产品来有效计算整齐近似微数。此外,我们为预测级概率的概率提供了一种近似差。我们观察到,当渔业信息矩阵中最小值接近于2美元正常化率时,近似于零。我们用TensorFlow实施的方法演示了该方法的演示,我们展示了基于预测性不确定性的图像的有意义的排序,对于两个基于勒Net的神经网络的概率概率概率概率概率来说是有约束的。 我们观察到,当渔业信息矩阵最小值接近值接近2美元时,只能用SMAIS和CI-10的测算数据测算出正确的数据。