Regenerating natural language explanations for science questions is a challenging task for evaluating complex multi-hop and abductive inference capabilities. In this setting, Transformers trained on human-annotated explanations achieve state-of-the-art performance when adopted as cross-encoder architectures. However, while much attention has been devoted to the quality of the constructed explanations, the problem of performing abductive inference at scale is still under-studied. As intrinsically not scalable, the cross-encoder architectural paradigm is not suitable for efficient multi-hop inference on massive facts banks. To maximise both accuracy and inference time, we propose a hybrid abductive solver that autoregressively combines a dense bi-encoder with a sparse model of explanatory power, computed leveraging explicit patterns in the explanations. Our experiments demonstrate that the proposed framework can achieve performance comparable with the state-of-the-art cross-encoder while being $\approx 50$ times faster and scalable to corpora of millions of facts. Moreover, we study the impact of the hybridisation on semantic drift and science question answering without additional training, showing that it boosts the quality of the explanations and contributes to improved downstream inference performance.
翻译:对科学问题进行自然语言的再生自然解释是评估复杂的多重机会和诱拐推断能力的一项艰巨任务。在这一背景下,受过人类附加说明解释培训的变异器在被采用为交叉编码结构时达到最先进的性能。然而,虽然人们已经非常关注所设计的解释质量,但规模化的绑架推断问题仍然研究不足。由于本质上无法伸缩,交叉编码建筑范式不适合在大量事实库中进行有效的多重机会推断。为了尽量提高准确性和推断时间,我们建议一种混合式的绑架解答器,自动递增地将密集的双编码与稀少的解释能力模型结合起来,计算解释中的明确模式。我们的实验表明,拟议的框架的性能可以与最先进的交叉编码相比,而其价值为50美元,而且对数百万事实库来说是可缩放的。此外,我们研究混合化对精度流和科学问题解释的影响,不需通过额外的培训来推动下游分析质量解释。我们实验表明,拟议框架的性能与最接近于最先进的交叉编码。