Our goal, in the context of open-domain textual question-answering (QA), is to explain answers by showing the line of reasoning from what is known to the answer, rather than simply showing a fragment of textual evidence (a "rationale'"). If this could be done, new opportunities for understanding and debugging the system's reasoning become possible. Our approach is to generate explanations in the form of entailment trees, namely a tree of multipremise entailment steps from facts that are known, through intermediate conclusions, to the hypothesis of interest (namely the question + answer). To train a model with this skill, we created ENTAILMENTBANK, the first dataset to contain multistep entailment trees. Given a hypothesis (question + answer), we define three increasingly difficult explanation tasks: generate a valid entailment tree given (a) all relevant sentences (b) all relevant and some irrelevant sentences, or (c) a corpus. We show that a strong language model can partially solve these tasks, in particular when the relevant sentences are included in the input (e.g., 35% of trees for (a) are perfect), and with indications of generalization to other domains. This work is significant as it provides a new type of dataset (multistep entailments) and baselines, offering a new avenue for the community to generate richer, more systematic explanations.
翻译:我们的目标是,在开放式文本解答问题的背景下,通过展示答案所知道的逻辑线来解释答案,而不是简单地展示文本证据的片段(一个“解释性” ) 。如果能够做到这一点,理解和调试系统推理的新机会就会成为可能。我们的方法是以包含树的形式作出解释,即从已知的事实(通过中间结论)到感兴趣的假设(即问题+答案),多层树意味着步骤。为了用这种技巧来训练一个模型,我们创建了第一个包含多级要求树的数据集。根据假设(问题+回答),我们定义了三个日益困难的解释任务:产生一个有效的要求树(a) 所有有关的句子(b) 所有相关和一些无关的句子,或(c) 材料。我们表明,一个强有力的语言模式可以部分地解决这些任务,特别是当输入相关句子时(例如,35 %的树是包含多级要求树的模型),这是一个更精确的路径(a) 提供更精细的模板。