Epistemic Uncertainty is a measure of the lack of knowledge of a learner which diminishes with more evidence. While existing work focuses on using the variance of the Bayesian posterior due to parameter uncertainty as a measure of epistemic uncertainty, we argue that this does not capture the part of lack of knowledge induced by model misspecification. We discuss how the excess risk, which is the gap between the generalization error of a predictor and the Bayes predictor, is a sound measure of epistemic uncertainty which captures the effect of model misspecification. We thus propose a principled framework for directly estimating the excess risk by learning a secondary predictor for the generalization error and subtracting an estimate of aleatoric uncertainty, i.e., intrinsic unpredictability. We discuss the merits of this novel measure of epistemic uncertainty, and highlight how it differs from variance-based measures of epistemic uncertainty and addresses its major pitfall. Our framework, Direct Epistemic Uncertainty Prediction (DEUP) is particularly interesting in interactive learning environments, where the learner is allowed to acquire novel examples in each round. Through a wide set of experiments, we illustrate how existing methods in sequential model optimization can be improved with epistemic uncertainty estimates from DEUP, and how DEUP can be used to drive exploration in reinforcement learning. We also evaluate the quality of uncertainty estimates from DEUP for probabilistic image classification and predicting synergies of drug combinations.
翻译:不确定性是衡量学习者缺乏知觉的一种衡量方法,这种知觉会减少更多的证据。虽然现有工作侧重于使用贝叶西亚后方的变异性,因为参数不确定性是分辨不确定性的一种衡量方法,但我们认为,这并不能反映模型偏差导致的知识缺乏的部分。我们讨论了超风险,即预测者与拜斯预测者预测者之间一般化错误之间的差值,是反映模型误差效应的典型不确定性的正确度度度度度。因此,我们提出了一个原则框架,用于直接估计过重风险,为此,我们学习了通用错误的二次预测,并减去了测深层不确定性的估计数,即内在不可预测性的估计数。我们讨论了这种新颖的隐喻不确定性度的优点,并着重指出了这种超重风险与基于差异的误差度的测算方法有何不同之处。我们的框架,即直辨误差的不确定性预测,它反映了模型的误差效应。我们特别有趣的是互动式学习环境环境,在这个环境中,学习了泛度预测错误的预估误差性预测,并减去了测结果,如何在每一回合中,DEREer能够用新的评估方法来模拟。