Recently neural response generation models have leveraged large pre-trained transformer models and knowledge snippets to generate relevant and informative responses. However, this does not guarantee that generated responses are factually correct. In this paper, we examine factual correctness in knowledge-grounded neural response generation models. We present a human annotation setup to identify three different response types: responses that are factually consistent with respect to the input knowledge, responses that contain hallucinated knowledge, and non-verifiable chitchat style responses. We use this setup to annotate responses generated using different stateof-the-art models, knowledge snippets, and decoding strategies. In addition, to facilitate the development of a factual consistency detector, we automatically create a new corpus called Conv-FEVER that is adapted from the Wizard of Wikipedia dataset and includes factually consistent and inconsistent responses. We demonstrate the benefit of our Conv-FEVER dataset by showing that the models trained on this data perform reasonably well to detect factually inconsistent responses with respect to the provided knowledge through evaluation on our human annotated data. We will release the Conv-FEVER dataset and the human annotated responses.
翻译:最近神经反应生成模型利用了经过事先培训的大型变压器模型和知识片段来生成相关且信息丰富的响应。然而,这并不能保证生成的响应在事实上是正确的。在本文件中,我们审视了基于知识的神经反应生成模型中的事实正确性。我们提出了一个人类说明设置,以确定三种不同的响应类型:与输入知识相一致的响应,包含致幻知识的响应,以及非可核实的切片风格响应。我们用这个设置来对使用不同最新模型、知识片段和解码战略生成的注释性响应进行说明。此外,为了便利开发一个事实一致性探测器,我们自动创建了一套名为Conv-FEffele的新材料,根据维基百科数据集向导师的数据集进行调整,并包含事实上一致和不一致的响应。我们通过显示在这些数据上接受培训的模型在评估我们人类附加注释的数据时对所提供的知识进行事实上不一致的响应,来证明我们Con-FEver数据集的好处。我们将发布Con-FEWER数据集和人类附加说明的答复。