Explaining the predictions made by complex machine learning models helps users to understand and accept the predicted outputs with confidence. One promising way is to use similarity-based explanation that provides similar instances as evidence to support model predictions. Several relevance metrics are used for this purpose. In this study, we investigated relevance metrics that can provide reasonable explanations to users. Specifically, we adopted three tests to evaluate whether the relevance metrics satisfy the minimal requirements for similarity-based explanation. Our experiments revealed that the cosine similarity of the gradients of the loss performs best, which would be a recommended choice in practice. In addition, we showed that some metrics perform poorly in our tests and analyzed the reasons of their failure. We expect our insights to help practitioners in selecting appropriate relevance metrics and also aid further researches for designing better relevance metrics for explanations.
翻译:解释复杂的机器学习模型所作的预测,有助于用户有信心地理解和接受预测产出; 一种有希望的方法是使用基于相似情况的类似解释,作为支持模型预测的证据; 为此使用了若干相关指标; 在这项研究中,我们调查了能够向用户提供合理解释的相关指标; 具体地说,我们通过了三项测试,以评估相关指标是否符合基于相似性的解释的最起码要求; 我们的实验显示,损失梯度的相近性表现最佳,这是实践中推荐的一种选择。 此外,我们表明,一些指标在我们的测试中表现不佳,并分析了失败的原因。 我们期望我们的洞察力有助于从业人员选择适当的相关指标,并有助于进一步研究设计更好的解释相关性指标。