Crucial for building trust in deep learning models for critical real-world applications is efficient and theoretically sound uncertainty quantification, a task that continues to be challenging. Useful uncertainty information is expected to have two key properties: It should be valid (guaranteeing coverage) and discriminative (more uncertain when the expected risk is high). Moreover, when combined with deep learning (DL) methods, it should be scalable and affect the DL model performance minimally. Most existing Bayesian methods lack frequentist coverage guarantees and usually affect model performance. The few available frequentist methods are rarely discriminative and/or violate coverage guarantees due to unrealistic assumptions. Moreover, many methods are expensive or require substantial modifications to the base neural network. Building upon recent advances in conformal prediction and leveraging the classical idea of kernel regression, we propose Locally Valid and Discriminative confidence intervals (LVD), a simple, efficient and lightweight method to construct discriminative confidence intervals (CIs) for almost any DL model. With no assumptions on the data distribution, such CIs also offer finite-sample local coverage guarantees (contrasted to the simpler marginal coverage). Using a diverse set of datasets, we empirically verify that besides being the only locally valid method, LVD also exceeds or matches the performance (including coverage rate and prediction accuracy) of existing uncertainty quantification methods, while offering additional benefits in scalability and flexibility.
翻译:对关键现实应用的深层次学习模型建立信任的关键在于高效和理论上可靠的不确定性量化,这一任务在理论上仍然具有挑战性。有用的不确定性信息预计将有两个关键属性:有效(保证覆盖面)和歧视性(预期风险高时更不确定),此外,如果与深层次学习(DL)方法相结合,它应当可扩展并影响DL模型的性能,多数现有的巴伊西亚方法缺乏频繁的覆盖保障,通常影响模型性能。由于很少的现有常住方法很少具有歧视性和/或由于不现实的假设而违反覆盖保障。此外,许多方法费用昂贵,或需要对基础神经网络进行重大修改。根据近期在合规预测和利用典型的内核回归概念方面取得的进展,我们提出本地有效和差异性信任间隔(LVD),这是为几乎所有DL模型构建歧视性信任间隔(CIs)的简单、高效和轻度方法。在数据分布上没有假设,这类中心还提供有限的局部本地覆盖保障(与边际覆盖相比更精确性),并且仅提供更精确的本地的准确性,同时提供我们现有数据评估的可靠率。