Reasoning over natural language is a long-standing goal for the research community. However, studies have shown that existing language models are inadequate in reasoning. To address the issue, we present POET, a novel reasoning pre-training paradigm. Through pre-training language models with programs and their execution results, POET empowers language models to harvest the reasoning knowledge possessed by program executors via a data-driven approach. POET is conceptually simple and can be instantiated by different kinds of program executors. In this paper, we showcase two simple instances POET-Math and POET-Logic, in addition to a complex instance, POET-SQL. Experimental results on six benchmarks demonstrate that POET can significantly boost model performance in natural language reasoning, such as numerical reasoning, logical reasoning, and multi-hop reasoning. POET opens a new gate on reasoning-enhancement pre-training, and we hope our analysis would shed light on the future research of reasoning like program executors.
翻译:自然语言是研究界的长期目标。然而,研究显示,现有的语言模式在推理方面是不充分的。为了解决这一问题,我们介绍了一个新的培训前推理范式POET。通过使用程序及其执行结果的培训前语言模型,POET赋予语言模型权力,通过数据驱动的方法获取程序执行者掌握的推理知识。POET在概念上是简单的,可以由不同种类的方案执行者即时进行。在本文中,我们展示了两个简单的POET-Math和POET-Logic案例,此外还有复杂的实例,POET-SQL。 六个基准的实验结果表明,POET能够极大地提高自然语言推理的模型性能,如数字推理、逻辑推理和多动推理。POET打开了推理前培训的新大门,我们希望我们的分析能够揭示未来对像程序执行者一样的推理学的研究。