Crucial for building trust in deep learning models for critical real-world applications is efficient and theoretically sound uncertainty quantification, a task that continues to be challenging. Useful uncertainty information is expected to have two key properties: It should be valid (guaranteeing coverage) and discriminative (more uncertain when the expected risk is high). Moreover, when combined with deep learning (DL) methods, it should be scalable and affect the DL model performance minimally. Most existing Bayesian methods lack frequentist coverage guarantees and usually affect model performance. The few available frequentist methods are rarely discriminative and/or violate coverage guarantees due to unrealistic assumptions. Moreover, many methods are expensive or require substantial modifications to the base neural network. Building upon recent advances in conformal prediction [13, 33] and leveraging the classical idea of kernel regression, we propose Locally Valid and Discriminative prediction intervals (LVD), a simple, efficient, and lightweight method to construct discriminative prediction intervals (PIs) for almost any DL model. With no assumptions on the data distribution, such PIs also offer finite-sample local coverage guarantees (contrasted to the simpler marginal coverage). We empirically verify, using diverse datasets, that besides being the only locally valid method for DL, LVD also exceeds or matches the performance (including coverage rate and prediction accuracy) of existing uncertainty quantification methods, while offering additional benefits in scalability and flexibility.
翻译:对关键现实应用的深层学习模型建立信任的关键在于高效和理论上可靠的不确定性量化,这一任务在理论上仍然具有挑战性。 有用的不确定性信息预计将有两个关键属性:有效(保证覆盖面)和歧视性(预期风险高时更不确定),此外,当与深层学习(DL)方法相结合时,它应当可缩放,对DL模型的性能影响极小。多数现有的巴伊西亚方法缺乏经常性覆盖保障,通常会影响模型性能。很少有常见方法很少具有歧视性和/或由于不现实的假设而破坏覆盖保障。此外,许多方法费用昂贵,或需要对基础神经网络进行重大修改:它应当有效(保证覆盖面的保障)和具有歧视性(当预期风险很高时,我们提出局部有效和模糊的预测间隔,这是为几乎所有DL模型构建歧视性预测间隔的简单、高效和轻量的方法。在数据分布上没有假设,这种PIS还提供有限的局部本地覆盖保障,同时使用简单的数据比值(我们仅提供实地预测率的精确度,同时提供实地比标准比标准) 。