The application of machine learning based decision making systems in safety critical areas requires reliable high certainty predictions. Reject options are a common way of ensuring a sufficiently high certainty of predictions made by the system. While being able to reject uncertain samples is important, it is also of importance to be able to explain why a particular sample was rejected. However, explaining general reject options is still an open problem. We propose a model agnostic method for locally explaining arbitrary reject options by means of interpretable models and counterfactual explanations.
翻译:在安全关键地区应用基于机器学习的决策系统需要可靠的高度确定性预测。拒绝选项是确保系统预测具有足够高度确定性的常见方法。虽然能够拒绝不确定的样本很重要,但解释为什么拒绝特定样本也很重要。然而,解释一般性拒绝选项仍然是一个尚未解决的问题。我们提出了一个示范不可知性方法,通过可解释的模式和反事实解释,在当地解释任意拒绝选项。