Seq2seq models based on Recurrent Neural Networks (RNNs) have recently received a lot of attention in the domain of Semantic Parsing for Question Answering. While in principle they can be trained directly on pairs (natural language utterances, logical forms), their performance is limited by the amount of available data. To alleviate this problem, we propose to exploit various sources of prior knowledge: the well-formedness of the logical forms is modeled by a weighted context-free grammar; the likelihood that certain entities present in the input utterance are also present in the logical form is modeled by weighted finite-state automata. The grammar and automata are combined together through an efficient intersection algorithm to form a soft guide ("background") to the RNN. We test our method on an extension of the Overnight dataset and show that it not only strongly improves over an RNN baseline, but also outperforms non-RNN models based on rich sets of hand-crafted features.
翻译:基于经常性神经网络的Seq2seq 模型(Seq2sqeq 模型)最近在Sementic Parring for question Regular Exproductions (RNN) 领域受到大量关注。 虽然原则上他们可以直接在对子上接受培训( 自然语言语句、 逻辑形式), 但其性能受可用数据数量的限制。 为了缓解这一问题, 我们提议利用各种先前知识来源: 逻辑形式的完善性是由加权的无上下文语法模拟的; 输入语句中的某些实体也以逻辑形式存在的可能性也由加权的定数自动式模型模拟。 语法和自动模型通过高效的交叉算法结合在一起, 形成一个软性指南(“ 背地 ”) 至 RNNN 。 我们测试我们的方法, 延长了超夜数据集, 并表明它不仅大大改进了 RNN 基线, 而且还超越了基于丰富手制特征的非 RNNN 模型。