Prior to deep learning the semantic parsing community has been interested in understanding and modeling the range of possible word alignments between natural language sentences and their corresponding meaning representations. Sequence-to-sequence models changed the research landscape suggesting that we no longer need to worry about alignments since they can be learned automatically by means of an attention mechanism. More recently, researchers have started to question such premise. In this work we investigate whether seq2seq models can handle both simple and complex alignments. To answer this question we augment the popular Geo semantic parsing dataset with alignment annotations and create Geo-Aligned. We then study the performance of standard seq2seq models on the examples that can be aligned monotonically versus examples that require more complex alignments. Our empirical study shows that performance is significantly better over monotonic alignments.
翻译:在深入学习语义分析社区之前,我们一直有兴趣理解和模拟自然语言句及其相应含义表达法之间可能的词汇匹配范围。 顺序对顺序模型改变了研究场景,表明我们不再需要担心对齐,因为它们可以通过关注机制自动学习。 最近,研究人员开始质疑这种前提。 在这项工作中,我们调查后继2当量模型能否同时处理简单和复杂的对齐问题。 为了回答这个问题,我们用对齐说明来提升流行的地理语义对齐值数据集,并创建Geo- Along。 然后我们研究标准后继2当量模型的性能,这些模型可以单调,而实例则需要更加复杂的对齐。我们的经验研究表明,相对于单声比对齐值的匹配性能要好得多。