Pre-trained language models have made great progress on dialogue tasks. However, these models are typically trained on surface dialogue text, thus are proven to be weak in understanding the main semantic meaning of a dialogue context. We investigate Abstract Meaning Representation (AMR) as explicit semantic knowledge for pre-training models to capture the core semantic information in dialogues during pre-training. In particular, we propose a semantic-based pre-training framework that extends the standard pre-training framework (Devlin et al., 2019) by three tasks for learning 1) core semantic units, 2) semantic relations and 3) the overall semantic representation according to AMR graphs. Experiments on the understanding of both chit-chats and task-oriented dialogues show the superiority of our model. To our knowledge, we are the first to leverage a deep semantic representation for dialogue pre-training.
翻译:培训前语言模式在对话任务方面取得了巨大进展,但是,这些模式通常在表面对话文本方面受过培训,因此在理解对话背景的主要语义含义方面证明是薄弱的。我们把“抽象含义”作为培训前模式的明确语义知识来调查培训前模式,以获取培训前对话中的核心语义信息。我们特别提议了一个基于语义的训练前框架,通过三个学习任务来扩展标准培训前框架(Devlin等人,2019年),即学习1个核心语义单元,2个语义关系和3个根据AMR图表显示的总体语义代表。关于理解语义和面向任务的对话的实验显示了我们模式的优越性。就我们的知识而言,我们是第一个利用深层语义代表来展开培训前对话的。