Whereas the ability of deep networks to produce useful predictions on many kinds of data has been amply demonstrated, estimating the reliability of these predictions remains challenging. Sampling approaches such as MC-Dropout and Deep Ensembles have emerged as the most popular ones for this purpose. Unfortunately, they require many forward passes at inference time, which slows them down. Sampling-free approaches can be faster but suffer from other drawbacks, such as lower reliability of uncertainty estimates, difficulty of use, and limited applicability to different types of tasks and data. In this work, we introduce a sampling-free approach that is generic and easy to deploy, while producing reliable uncertainty estimates on par with state-of-the-art methods at a significantly lower computational cost. It is predicated on training the network to produce the same output with and without additional information about that output. At inference time, when no prior information is given, we use the network's own prediction as the additional information. We prove that the difference between the two predictions is an accurate uncertainty estimate and demonstrate our approach on various types of tasks and applications.
翻译:虽然深度网络对许多种类的数据作出有用预测的能力已经得到充分证明,但估计这些预测的可靠性仍然具有挑战性。诸如MC-Dropout和深团等抽样方法已成为最受欢迎的方法。不幸的是,它们要求许多在推论时间提前传递,从而放慢了速度。无抽样方法可以更快,但也有其他缺点,例如不确定性估计的可靠性较低、使用困难以及对不同类型任务和数据的有限适用性。在这项工作中,我们采用了一种通用的、易于使用的无抽样方法,同时在计算成本低得多的情况下得出与最新方法相同的可靠不确定性估计数。这些方法的前提是对网络进行培训,以便用和不增加关于该产出的信息来产生同样的产出。在推断期间,当没有提供先前信息时,我们用网络自己的预测作为补充信息。我们证明,两种预测之间的差异是一种准确的不确定性估计,并表明我们对于各种任务和应用的方法。