Knowledge graph (KG) question generation (QG) aims to generate natural language questions from KGs and target answers. Previous works mostly focus on a simple setting which is to generate questions from a single KG triple. In this work, we focus on a more realistic setting where we aim to generate questions from a KG subgraph and target answers. In addition, most of previous works built on either RNN-based or Transformer based models to encode a linearized KG sugraph, which totally discards the explicit structure information of a KG subgraph. To address this issue, we propose to apply a bidirectional Graph2Seq model to encode the KG subgraph. Furthermore, we enhance our RNN decoder with node-level copying mechanism to allow directly copying node attributes from the KG subgraph to the output question. Both automatic and human evaluation results demonstrate that our model achieves new state-of-the-art scores, outperforming existing methods by a significant margin on two QG benchmarks. Experimental results also show that our QG model can consistently benefit the Question Answering (QA) task as a mean of data augmentation.
翻译:知识图谱问答生成旨在从知识图谱和目标答案中生成自然语言问题。之前的工作大多集中于简单的情景,即从单个知识图谱三元组生成问题。本工作着眼于更现实的情景,即从知识图谱子图和目标答案生成问题。此外,之前的大多数工作都是基于基于RNN或Transformer的模型对线性化的知识图谱子图进行编码,这完全丢弃了知识图谱子图的显式结构信息。为了解决这个问题,我们提出了一个双向Graph2Seq模型来编码知识图谱子图。此外,我们通过节点级别的复制机制增强了RNN解码器,允许直接从知识图谱子图将节点属性复制到输出问题中。自动评估和人工评估结果均表明,我们的模型实现了新的最高分数,在两个问答生成基准测试中都比现有方法表现更好。实验结果还表明,我们的问答生成模型作为数据增强的一种手段可以持续地使问答任务受益。