A crucial requirement for reliable deployment of deep learning models for safety-critical applications is the ability to identify out-of-distribution (OOD) data points, samples which differ from the training data and on which a model might underperform. Previous work has attempted to tackle this problem using uncertainty estimation techniques. However, there is empirical evidence that a large family of these techniques do not detect OOD reliably in classification tasks. This paper gives a theoretical explanation for said experimental findings and illustrates it on synthetic data. We prove that such techniques are not able to reliably identify OOD samples in a classification setting, since their level of confidence is generalized to unseen areas of the feature space. This result stems from the interplay between the representation of ReLU networks as piece-wise affine transformations, the saturating nature of activation functions like softmax, and the most widely-used uncertainty metrics.
翻译:可靠地部署安全关键应用的深层学习模型的关键要求是能够确定分布外(OOD)数据点,这些样本不同于培训数据,而且模型可能不完善。以前的工作曾试图利用不确定性估计技术解决这一问题。然而,有经验证据表明,这些技术中的一大批在分类任务中无法可靠地检测OOD。本文从理论上解释了上述实验结果,并用合成数据加以说明。我们证明,这些技术无法在分类环境中可靠地识别OOD样本,因为其信任度被普遍推广到地物空间的隐蔽区域。这一结果源于ReLU网络作为片形形形形形形形形形形形形形形变、软轴等活性功能的饱和性质以及最广泛使用的不确定度度值之间的相互作用。