Trust is a crucial factor affecting the adoption of machine learning (ML) models. Qualitative studies have revealed that end-users, particularly in the medical domain, need models that can express their uncertainty in decision-making allowing users to know when to ignore the model's recommendations. However, existing approaches for quantifying decision-making uncertainty are not model-agnostic, or they rely on complex statistical derivations that are not easily understood by laypersons or end-users, making them less useful for explaining the model's decision-making process. This work proposes a set of class-independent meta-heuristics that can characterize the complexity of an instance in terms of factors are mutually relevant to both human and ML decision-making. The measures are integrated into a meta-learning framework that estimates the risk of misclassification. The proposed framework outperformed predicted probabilities in identifying instances at risk of being misclassified. The proposed measures and framework hold promise for improving model development for more complex instances, as well as providing a new means of model abstention and explanation.
翻译:信任是影响机器学习(ML)模型采纳的关键因素。定性研究已经表明,终端用户,特别是在医疗领域,需要能够表达其决策的不确定性的模型,以便用户知道何时忽略模型的推荐。然而,现有的量化决策不确定性的方法不是针对模型的,或者依赖于复杂的统计推导,这些推导对于普通人或终端用户来说不容易理解,使其不适于解释模型的决策过程。本研究提出了一组独立于类别的元启发式,可以从人类和机器学习决策方面相互关联的因素的角度来表征实例的复杂性。这些测量值被集成到一种元学习框架中,该框架估计了误分类的风险。提出的框架在识别处于被误分类风险的实例方面优于预测概率。提出的测量值和框架有望改进用于更复杂实例的模型开发,同时为解释和拟除模型提供了一种新的手段。