Summarization of long sequences into a concise statement is a core problem in natural language processing, requiring non-trivial understanding of the input. Based on the promising results of graph neural networks on highly structured data, we develop a framework to extend existing sequence encoders with a graph component that can reason about long-distance relationships in weakly structured data such as text. In an extensive evaluation, we show that the resulting hybrid sequence-graph models outperform both pure sequence models as well as pure graph models on a range of summarization tasks.
翻译:将长序列归纳为简明扼要的语句是自然语言处理的一个核心问题,需要非三思而行地理解输入。基于高结构数据的图形神经网络的可喜结果,我们开发了一个框架,以扩展现有的序列编码器,其中含有一个图形部分,在像文本这样的结构薄弱的数据中可以解释长距离关系。在一项广泛的评估中,我们表明由此产生的混合序列绘图模型既优于纯序列模型,也优于一系列总和任务的纯图形模型。