Traditional automated theorem provers have relied on manually tuned heuristics to guide how they perform proof search. Recently, however, there has been a surge of interest in the design of learning mechanisms that can be integrated into theorem provers to improve their performance automatically. In this work, we introduce TRAIL, a deep learning-based approach to theorem proving that characterizes core elements of saturation-based theorem proving within a neural framework. TRAIL leverages (a) an effective graph neural network for representing logical formulas, (b) a novel neural representation of the state of a saturation-based theorem prover in terms of processed clauses and available actions, and (c) a novel representation of the inference selection process as an attention-based action policy. We show through a systematic analysis that these components allow TRAIL to significantly outperform previous reinforcement learning-based theorem provers on two standard benchmark datasets (up to 36% more theorems proved). In addition, to the best of our knowledge, TRAIL is the first reinforcement learning-based approach to exceed the performance of a state-of-the-art traditional theorem prover on a standard theorem proving benchmark (solving up to 17% more problems).
翻译:传统自动理论验证器依靠人工调整的神经系统来指导他们如何进行证据搜索。然而,最近,人们对设计学习机制的兴趣激增,可以将其纳入理论验证器,从而自动改进其性能。在这项工作中,我们引入了TRAIL, 这是一种深层次的基于学习的理论验证器,用以证明在神经框架内以饱和为基础的理论验证的核心要素的特点。TRAIL的杠杆(a) 代表逻辑公式的有效图形神经网络,(b) 在处理的条款和现有行动方面,基于饱和的理论验证器状态的新神经系统,以及(c) 将推断选择过程作为注重行动政策的新表述。我们通过系统分析表明,这些要素使得TRAIL在两个标准基准数据集上大大超越了先前强化的基于理论的理论验证器(最多为36%)。此外,除了我们最先进的知识外,TRAIL是第一个以强化学习为基础的方法,用以证明一种以超过标准标准的标准的17问题。我们通过系统分析表明,TRAIL能够大大改进以前的强化基于学习的理论验证方法。