Unbiased assessment of the predictivity of models learnt by supervised machine-learning methods requires knowledge of the learned function over a reserved test set (not used by the learning algorithm). The quality of the assessment depends, naturally, on the properties of the test set and on the error statistic used to estimate the prediction error. In this work we tackle both issues, proposing a new predictivity criterion that carefully weights the individual observed errors to obtain a global error estimate, and using incremental experimental design methods to "optimally" select the test points on which the criterion is computed. Several incremental constructions are studied, including greedy-packing (coffee-house design), support points and kernel herding techniques. Our results show that the incremental and weighted versions of the latter two, based on Maximum Mean Discrepancy concepts, yield superior performance. An industrial test case provided by the historical French electricity supplier (EDF) illustrates the practical relevance of the methodology, indicating that it is an efficient alternative to expensive cross-validation techniques.
翻译:在这项工作中,我们处理这两个问题,提出一个新的预测性标准,仔细权衡个人观察到的错误,以获得全球误差估计,并使用渐进的实验设计方法“最理想地”选择计算标准所根据的测试点。研究了一些渐进式的构造,包括贪婪包装(餐饮设计)、支持点和内嵌套圈技术。我们的结果显示,后两种方法的递增和加权版本基于最大平均值差异概念,产生优异性能。法国历史电力供应商(EDF)提供的一个工业试验案例说明了该方法的实际相关性,表明它是昂贵的交叉验证技术的一种有效替代方法。