A growing body of work studies how to answer a question or verify a claim by generating a natural language "proof": a chain of deductive inferences yielding the answer based on a set of premises. However, these methods can only make sound deductions when they follow from evidence that is given. We propose a new system that can handle the underspecified setting where not all premises are stated at the outset; that is, additional assumptions need to be materialized to prove a claim. By using a natural language generation model to abductively infer a premise given another premise and a conclusion, we can impute missing pieces of evidence needed for the conclusion to be true. Our system searches over two fringes in a bidirectional fashion, interleaving deductive (forward-chaining) and abductive (backward-chaining) generation steps. We sample multiple possible outputs for each step to achieve coverage of the search space, at the same time ensuring correctness by filtering low-quality generations with a round-trip validation procedure. Results on a modified version of the EntailmentBank dataset and a new dataset called Everyday Norms: Why Not? show that abductive generation with validation can recover premises across in- and out-of-domain settings.
翻译:越来越多的工作研究如何解答问题或通过产生自然语言“可靠”来核实索赔要求:一连串推理推理推论,根据一套前提得出答案。然而,这些方法只能从所提供的证据中作出合理的扣减。我们建议了一个新的系统,可以处理未事先说明所有前提的未指定背景;也就是说,额外的假设需要实现才能证明索赔要求。通过使用一种自然语言生成模型,将一个前提和结论的前提随机地推入一个前提,我们可以对结论真实所需的缺失证据进行估算。我们的系统以双向方式对两个边缘进行搜索,中间的推理(前向链)和绑架(后向链)生成步骤。我们为每个步骤取样了多种可能的产出,以覆盖搜索空间,同时用一个圆程验证程序过滤低质量的世代,以确保正确性。关于修改后版的Bank数据集和称为每天标准的新数据集的结果:为什么不恢复?在每一代制式规范中显示该绑架期的校正。