In order to achieve deep natural language understanding, syntactic constituent parsing is a vital step, highly demanded by many artificial intelligence systems to process both text and speech. One of the most recent proposals is the use of standard sequence-to-sequence models to perform constituent parsing as a machine translation task, instead of applying task-specific parsers. While they show a competitive performance, these text-to-parse transducers are still lagging behind classic techniques in terms of accuracy, coverage and speed. To close the gap, we here extend the framework of sequence-to-sequence models for constituent parsing, not only by providing a more powerful neural architecture for improving their performance, but also by enlarging their coverage to handle the most complex syntactic phenomena: discontinuous structures. To that end, we design several novel linearizations that can fully produce discontinuities and, for the first time, we test a sequence-to-sequence model on the main discontinuous benchmarks, obtaining competitive results on par with task-specific discontinuous constituent parsers and achieving state-of-the-art scores on the (discontinuous) English Penn Treebank.
翻译:为了实现深入的自然语言理解,合成构件剖析是一个关键步骤,许多人工智能系统对处理文本和言语要求很高。最近的一项提议是使用标准序列到顺序模型,将构成剖析作为一种机器翻译任务,而不是应用特定任务剖析员。虽然这些文本到分类的转换员表现出了竞争性的性能,但在准确性、覆盖面和速度方面仍然落后于经典技术。为了缩小这一差距,我们在此扩展了成份分割的顺序到顺序模型框架,不仅通过提供更强大的神经结构来改进其性能,而且还通过扩大覆盖范围来处理最复杂的合成现象:不连续结构。为此,我们设计了几种新的线性化,可以完全产生不连续性,并且首次在主要的不连续基准上测试一个序列到顺序的顺序模型,在与特定任务不连续的构件分离器相匹配时取得竞争性结果,并在(不连续的)英金库中实现状态分级。