Modern deep learning based classifiers show very high accuracy on test data but this does not provide sufficient guarantees for safe deployment, especially in high-stake AI applications such as medical diagnosis. Usually, predictions are obtained without a reliable uncertainty estimate or a formal guarantee. Conformal prediction (CP) addresses these issues by using the classifier's probability estimates to predict confidence sets containing the true class with a user-specified probability. However, using CP as a separate processing step after training prevents the underlying model from adapting to the prediction of confidence sets. Thus, this paper explores strategies to differentiate through CP during training with the goal of training model with the conformal wrapper end-to-end. In our approach, conformal training (ConfTr), we specifically "simulate" conformalization on mini-batches during training. We show that CT outperforms state-of-the-art CP methods for classification by reducing the average confidence set size (inefficiency). Moreover, it allows to "shape" the confidence sets predicted at test time, which is difficult for standard CP. On experiments with several datasets, we show ConfTr can influence how inefficiency is distributed across classes, or guide the composition of confidence sets in terms of the included classes, while retaining the guarantees offered by CP.
翻译:现代深层学习分类显示,测试数据的准确性非常高,但这并不能为安全部署提供足够的保障,特别是在医疗诊断等高风险AI应用中。通常,在没有可靠的不确定性估计或正式保证的情况下获得预测。非正式预测(CP)通过使用分类者的概率估计来预测包含真实等级的置信套数来解决这些问题。然而,在培训后将CP作为一个单独的处理步骤,使基本模式无法适应信心套数的预测。因此,本文件探索了在培训期间通过CP进行区分的战略,培训的目的是用符合要求的包装终端到终端等培训模型。在我们的方法中,在培训期间,我们具体用“模拟”微型容器的符合性培训。我们显示CT与包含用户特定概率的真实类别相容。我们通过降低平均信任套数的大小(效率),将CP值的预测套数与标准CP值挂钩。在几个数据集的实验中,我们展示了ConfT构成的“模拟”性,同时将信任等级包含在各种等级中。