Most of the existing works for dialogue generation are data-driven models trained directly on corpora crawled from websites. They mainly focus on improving the model architecture to produce better responses but pay little attention to considering the quality of the training data contrastively. In this paper, we propose a multi-level contrastive learning paradigm to model the fine-grained quality of the responses with respect to the query. A Rank-aware Calibration (RC) network is designed to construct the multi-level contrastive optimization objectives. Since these objectives are calculated based on the sentence level, which may erroneously encourage/suppress the generation of uninformative/informative words. To tackle this incidental issue, on one hand, we design an exquisite token-level strategy for estimating the instance loss more accurately. On the other hand, we build a Knowledge Inference (KI) component to capture the keyword knowledge from the reference during training and exploit such information to encourage the generation of informative words. We evaluate the proposed model on a carefully annotated dialogue dataset and the results suggest that our model can generate more relevant and diverse responses compared to the baseline models.
翻译:对话生成的现有工作大多是直接在网站爬行的公司上培训的数据驱动模型,主要侧重于改进模型结构,以产生更好的反应,但很少注意对比地考虑培训数据的质量。在本文件中,我们建议采用多层次对比学习模式,以模拟对查询的精细反应质量。一个Rank-aware校准(RC)网络旨在构建多层次对比优化目标。由于这些目标是根据句子水平计算的,这可能错误地鼓励/压制生成非信息规范/信息化词汇。一方面,为了解决这一附带问题,我们设计了一个极具代表性的战略,以更准确地估计实例损失。另一方面,我们建立一个知识推理(KI)部分,从培训参考中获取关键词知识,并利用这种信息鼓励生成信息文字。我们评估了关于谨慎说明的对话数据集的拟议模型,结果表明我们的模型能够产生比基线模型更相关和多样化的反应。