Reliable application of machine learning-based decision systems in the wild is one of the major challenges currently investigated by the field. A large portion of established approaches aims to detect erroneous predictions by means of assigning confidence scores. This confidence may be obtained by either quantifying the model's predictive uncertainty, learning explicit scoring functions, or assessing whether the input is in line with the training distribution. Curiously, while these approaches all state to address the same eventual goal of detecting failures of a classifier upon real-life application, they currently constitute largely separated research fields with individual evaluation protocols, which either exclude a substantial part of relevant methods or ignore large parts of relevant failure sources. In this work, we systematically reveal current pitfalls caused by these inconsistencies and derive requirements for a holistic and realistic evaluation of failure detection. To demonstrate the relevance of this unified perspective, we present a large-scale empirical study for the first time enabling benchmarking confidence scoring functions w.r.t all relevant methods and failure sources. The revelation of a simple softmax response baseline as the overall best performing method underlines the drastic shortcomings of current evaluation in the abundance of publicized research on confidence scoring. Code and trained models are at https://github.com/IML-DKFZ/fd-shifts.
翻译:机器学习决策系统在实际应用中的可靠性是该领域目前研究的主要挑战之一。大部分已建立的方法旨在通过分配置信度来检测错误预测。这种信心可能是通过量化模型的预测不确定性、学习显式评分函数或评估输入是否符合训练分布等手段获得的。令人惊讶的是,虽然这些方法都宣称致力于在实际应用中检测分类器失败,但它们目前构成了大部分互相独立的研究领域,具有单独的评估协议,其中有些排除了相当一部分相关方法,有些则忽略了大量相关错误来源的数据。在这项工作中,我们系统地揭示了这些不一致性可能带来的当前问题,并得出了对于整体和实际评估失败检测的要求。为证明这种统一视角的重要性,我们首次进行了大规模实证研究,使评分函数针对所有相关方法和故障源进行基准测试。普通的softmax响应基线的揭示强调了当前置信度评估中的严重缺陷。代码和训练模型位于https://github.com/IML-DKFZ/fd-shifts。