Establishing retrieval-based dialogue systems that can select appropriate responses from the pre-built index has gained increasing attention from researchers. For this task, the adoption of pre-trained language models (such as BERT) has led to remarkable progress in a number of benchmarks. There exist two common approaches, including cross-encoders which perform full attention over the inputs, and bi-encoders that encode the context and response separately. The former gives considerable improvements in accuracy but is often inapplicable in practice for large-scale retrieval given the cost of the full attention required for each sample at test time. The latter is efficient for billions of indexes but suffers from sub-optimal performance. In this work, we propose to combine the best of both worlds to build a retrieval system. Specifically, we employ a fast bi-encoder to replace the traditional feature-based pre-retrieval model (such as BM25) and set the response re-ranking model as a more complicated architecture (such as cross-encoder). To further improve the effectiveness of our framework, we train the pre-retrieval model and the re-ranking model at the same time via mutual learning, which enables two models to learn from each other throughout the training process. We conduct experiments on two benchmarks and evaluation results demonstrate the efficiency and effectiveness of our proposed framework.
翻译:研究人员越来越重视建立检索对话系统,以便从预先设定的指数中选择适当的反应。对于这项任务,采用预先培训的语言模式(如BERT)已导致在若干基准方面取得显著进展。存在两种共同的方法,包括全面关注投入的交叉计算器,以及分别对背景和反应进行编码的双编码器。前者在准确性方面有很大改进,但鉴于每个样本在测试时需要充分注意的成本,后者往往无法用于大规模检索。对于数十亿指数而言,后者是有效的,但有亚最佳性能。在这项工作中,我们提议将两个世界的最佳组合起来,以建立一个检索系统。具体地说,我们使用快速双编码器取代传统的基于地貌的检索前模型(如BM25),并将反应的重新排序模式设置为更为复杂的结构(如交叉编码)。为了进一步提高我们框架的效力,我们培训了前检索模型和再排序模型,但有亚性性能。我们建议将这两个世界的最佳模型结合起来,以建立一个检索系统。具体地说,我们使用快速双编码器来取代传统的基于特征的检索前模型(例如BM25),将反应模型设置为较复杂的结构(例如交叉编码)。我们要),我们通过相互学习两个评价过程,然后从两个试验过程来学习。