An emerging line of research in Explainable NLP is the creation of datasets enriched with human-annotated explanations and rationales, used to build and evaluate models with step-wise inference and explanation generation capabilities. While human-annotated explanations are used as ground-truth for the inference, there is a lack of systematic assessment of their consistency and rigour. In an attempt to provide a critical quality assessment of Explanation Gold Standards (XGSs) for NLI, we propose a systematic annotation methodology, named Explanation Entailment Verification (EEV), to quantify the logical validity of human-annotated explanations. The application of EEV on three mainstream datasets reveals the surprising conclusion that a majority of the explanations, while appearing coherent on the surface, represent logically invalid arguments, ranging from being incomplete to containing clearly identifiable logical errors. This conclusion confirms that the inferential properties of explanations are still poorly formalised and understood, and that additional work on this line of research is necessary to improve the way Explanation Gold Standards are constructed.
翻译:在可解释的NLP中,正在出现的一项研究方针是建立以人类附加说明的解释和理由丰富的数据集,用于建立和评价具有逐步推论和解释生成能力的模型;虽然人类附加说明的解释被用作推论的地面真相,但缺乏对其一致性和严谨性的系统评估;为了对解释金标准(XGSs)进行严格的质量评估,我们提议了一个系统化的说明方法,名为解释零售核查(EEEV),以量化人附加说明的解释的逻辑有效性;EEV在三个主流数据集中的应用揭示出一个令人惊讶的结论,即大多数解释虽然在表面看起来一致,但代表了逻辑上无效的论据,从不完整到包含可明确识别的逻辑错误。这一结论证实解释的推断属性仍然不够正式化和理解,并且有必要就这一研究领域开展更多的工作,以改进解释黄金标准构建的方式。