Recent improvements in KG-to-text generation are due to additional auxiliary pre-training tasks designed to give the fine-tune task a boost in performance. These tasks require extensive computational resources while only suggesting marginal improvements. Here, we demonstrate that by fusing graph-aware elements into existing pre-trained language models, we are able to outperform state-of-the-art models and close the gap imposed by additional pre-training tasks. We do so by proposing a mask structure to capture neighborhood information and a novel type encoder that adds a bias to the graph-attention weights depending on the connection type. Experiments on two KG-to-text benchmark datasets show our models are competitive while involving fewer parameters and no additional pre-training tasks. By formulating the problem as a framework, we can interchange the various proposed components and begin interpreting KG-to-text generative models based on the topological and type information found in a graph.
翻译:KG- 文本生成的近期改进是由于增加了辅助性培训前任务,旨在提升微调任务的业绩。这些任务需要大量的计算资源,而只是提出微小的改进。在这里,我们证明,通过将图形觉悟元素嵌入现有的经培训前语言模型,我们能够超越最先进的模式,并缩小额外的培训前任务带来的差距。我们这样做的方式是提出一个遮罩结构,以捕捉邻里信息,并建立一个新型编码器,根据连接类型对图形注意权重增加偏差。对两个KG- 文本基准数据集的实验表明,我们的模型具有竞争力,同时涉及较少的参数,而没有额外的培训前任务。通过将问题作为一个框架,我们可以交换各种拟议组成部分,并开始根据图表中发现的表层和类型信息解释KG- 文本的基因模型。