Fact verification systems typically rely on neural network classifiers for veracity prediction which lack explainability. This paper proposes ProoFVer, which uses a seq2seq model to generate natural logic-based inferences as proofs. These proofs consist of lexical mutations between spans in the claim and the evidence retrieved, each marked with a natural logic operator. Claim veracity is determined solely based on the sequence of these operators. Hence, these proofs are faithful explanations, and this makes ProoFVer faithful by construction. Currently, ProoFVer has the highest label accuracy and the second-best Score in the FEVER leaderboard. Furthermore, it improves by 13.21% points over the next best model on a dataset with counterfactual instances, demonstrating its robustness. As explanations, the proofs show better overlap with human rationales than attention-based highlights and the proofs help humans predict model decisions correctly more often than using the evidence directly.
翻译:事实核查系统通常依赖神经网络分类来进行真实性预测,而这种预测缺乏解释性。本文建议采用ProoFVer, 使用一个后继2seq 模型来产生基于逻辑的自然推论作为证据。 这些证据包括索赔和检索的证据之间的逻辑突变,每个证据都有自然逻辑操作员的标记。 索赔的真实性完全依据这些操作员的顺序来确定。 因此,这些证据是忠实的解释, 使得ProoFVer在构建过程中忠实于他人。 目前, ProoFVer在FEWL 领头板上拥有最高标签准确性和第二高分。 此外,它比下一个以反事实实例显示其稳健性的最佳模型提高了13.21%。 作为解释, 证据显示与人的理由比关注重点和证据帮助人类正确预测模型决定比直接使用证据更加频繁。