Evaluating an explanation's faithfulness is desired for many reasons such as trust, interpretability and diagnosing the sources of model's errors. In this work, which focuses on the NLI task, we introduce the methodology of Faithfulness-through-Counterfactuals, which first generates a counterfactual hypothesis based on the logical predicates expressed in the explanation, and then evaluates if the model's prediction on the counterfactual is consistent with that expressed logic (i.e. if the new formula is \textit{logically satisfiable}). In contrast to existing approaches, this does not require any explanations for training a separate verification model. We first validate the efficacy of automatic counterfactual hypothesis generation, leveraging on the few-shot priming paradigm. Next, we show that our proposed metric distinguishes between human-model agreement and disagreement on new counterfactual input. In addition, we conduct a sensitivity analysis to validate that our metric is sensitive to unfaithful explanations.
翻译:出于信任、可解释性和分析模型错误来源等许多原因,评价解释的忠实性是可取的。在以NLI任务为重点的这项工作中,我们引入了信仰通过事实的方法,首先根据解释中表达的逻辑前提产生反事实假设,然后评估模型对反事实的预测是否与表达的逻辑一致(即,如果新公式为\textit{逻辑上可争议})。 与现有方法相反,这不需要为训练一个单独的核查模型作任何解释。我们首先验证自动反事实假设的产生效力,利用微弱的假象。接下来,我们表明我们拟议的指标区分了人类模型协议和对新的反事实投入的分歧。此外,我们进行了敏感性分析,以证实我们的计量对不真实的解释十分敏感。