Uncertainty quantification is at the core of the reliability and robustness of machine learning. In this paper, we provide a theoretical framework to dissect the uncertainty, especially the epistemic component, in deep learning into procedural variability (from the training procedure) and data variability (from the training data), which is the first such attempt in the literature to our best knowledge. We then propose two approaches to estimate these uncertainties, one based on influence function and one on batching. We demonstrate how our approaches overcome the computational difficulties in applying classical statistical methods. Experimental evaluations on multiple problem settings corroborate our theory and illustrate how our framework and estimation can provide direct guidance on modeling and data collection effort to improve deep learning performance.
翻译:不确定性量化是机器学习可靠性和稳健性的核心。在本文件中,我们提供了一个理论框架,在深入学习程序变异(从培训程序)和数据变异(从培训数据)方面,解析不确定性,特别是缩略语部分,这是文献中对我们最佳知识的首次尝试。然后,我们提出两种方法来估计这些不确定性,一种基于影响功能,另一种基于分批。我们展示了我们的方法如何克服在应用传统统计方法方面的计算困难。对多种问题设置的实验性评估证实了我们的理论,并说明了我们的框架和估计如何能为建模和数据收集工作提供直接指导,以改善深层学习绩效。