Predicting linearized Abstract Meaning Representation (AMR) graphs using pre-trained sequence-to-sequence Transformer models has recently led to large improvements on AMR parsing benchmarks. These parsers are simple and avoid explicit modeling of structure but lack desirable properties such as graph well-formedness guarantees or built-in graph-sentence alignments. In this work we explore the integration of general pre-trained sequence-to-sequence language models and a structure-aware transition-based approach. We depart from a pointer-based transition system and propose a simplified transition set, designed to better exploit pre-trained language models for structured fine-tuning. We also explore modeling the parser state within the pre-trained encoder-decoder architecture and different vocabulary strategies for the same purpose. We provide a detailed comparison with recent progress in AMR parsing and show that the proposed parser retains the desirable properties of previous transition-based approaches, while being simpler and reaching the new parsing state of the art for AMR 2.0, without the need for graph re-categorization.
翻译:使用经过事先训练的顺序至顺序变换模型的预测线性抽象表示图(AMR)最近导致对AMR分类基准的大幅改进,这些剖析器简单明了,避免对结构进行明确的建模,但缺乏适当的属性,如图表完善的保证或内嵌的图示感应调整。在这项工作中,我们探索了将一般经过训练的顺序至顺序的顺序语言模型和结构上意识到的过渡方法结合起来。我们偏离了基于点的过渡系统,提出了简化的过渡套件,旨在更好地利用经过训练的语文模型进行结构上的微调。我们还探索了在经过事先训练的编码解码器结构内建模的剖析器状态和用于同一目的的不同词汇战略。我们详细比较了AMR解析的最新进展,并表明拟议的剖析器保留了以前过渡方法的可取属性,同时简化并达到了对AMR 2.0的艺术新分解状态,而无需进行图表重新分类。