Dynamic neural networks are a recent technique that promises a remedy for the increasing size of modern deep learning models by dynamically adapting their computational cost to the difficulty of the inputs. In this way, the model can adjust to a limited computational budget. However, the poor quality of uncertainty estimates in deep learning models makes it difficult to distinguish between hard and easy samples. To address this challenge, we present a computationally efficient approach for post-hoc uncertainty quantification in dynamic neural networks. We show that adequately quantifying and accounting for both aleatoric and epistemic uncertainty through a probabilistic treatment of the last layers improves the predictive performance and aids decision-making when determining the computational budget. In the experiments, we show improvements on CIFAR-100, ImageNet, and Caltech-256 in terms of accuracy, capturing uncertainty, and calibration error.
翻译:----
动态神经网络是一种新的技术,通过动态调整其计算复杂度以适应输入数据难度,解决了当代深度学习模型越来越大的问题。然而,深度学习模型中不确定性估计的质量较差,使得难样本和易样本难以区分。为了应对这个挑战,我们提出了一种计算效率高的后处理不确定性量化方法,来处理动态神经网络的过度自信问题。通过对最后几层模型的概率处理,我们充分量化和考虑了aleatoric和epistemic不确定性,从而提高了预测性能,并在确定计算复杂度方面有帮助。实验证明,在CIFAR-100、ImageNet和Caltech-256数据集上,我们的方法在准确率、不确定性捕获和校准误差方面均有所提升。