Human conversations consist of reasonable and natural topic flows, which are observed as the shifts of the mentioned concepts across utterances. Previous chatbots that incorporate the external commonsense knowledge graph prove that modeling the concept shifts can effectively alleviate the dull and uninformative response dilemma. However, there still exists a gap between the concept relations in the natural conversation and those in the external commonsense knowledge graph, which is an issue to solve. Specifically, the concept relations in the external commonsense knowledge graph are not intuitively built from the conversational scenario but the world knowledge, which makes them insufficient for the chatbot construction. To bridge the above gap, we propose the method to supply more concept relations extracted from the conversational corpora and reconstruct an enhanced concept graph for the chatbot construction. In addition, we present a novel, powerful, and fast graph encoding architecture named the Edge-Transformer to replace the traditional GNN architecture. Experimental results on the Reddit conversation dataset indicate our proposed method significantly outperforms strong baseline systems and achieves new SOTA results. Further analysis individually proves the effectiveness of the enhanced concept graph and the Edge-Transformer architecture.
翻译:人类对话由合理和自然的话题流组成,这是上述概念跨语句的转变所观察到的。 包含外部常识知识图的前几个聊天机,证明模拟概念转变能够有效缓解无聊和不提供信息的应对困境。然而,自然对话的概念关系与外部常识知识图中的概念关系之间仍然存在差距,这是一个需要解决的问题。具体地说,外部常识知识图中的概念关系不是从谈话情景中直观地构建的,而是世界知识,这使得这些知识不足以构建聊天机。为了弥合上述差距,我们提议了方法,以提供从谈话公司中提取的更多概念关系,并为聊天机建设重建一个强化的概念图。此外,我们提出了一个名为Edge-Transent的新型、强大和快速图形编码结构,以取代传统的GNN结构。 Reddit对话数据集的实验结果表明,我们拟议的方法大大超越了强大的基线系统,并取得了新的SOTA结果。进一步的分析单独证明了强化的概念图表和Edge-Transfororum的实效。