Deep recommender systems jointly leverage the retrieval and ranking operations to generate the recommendation result. The retriever targets selecting a small set of relevant candidates from the entire items with high efficiency; while the ranker, usually more precise but time-consuming, is supposed to identify the best items out of the retrieved candidates with high precision. However, the retriever and ranker are usually trained in poorly-cooperative ways, leading to limited recommendation performances when working as an entirety. In this work, we propose a novel DRS training framework CoRR(short for Cooperative Retriever and Ranker), where the retriever and ranker can be mutually reinforced. On one hand, the retriever is learned from recommendation data and the ranker via knowledge distillation; knowing that the ranker is more precise, the knowledge distillation may provide extra weak-supervision signals for the improvement of retrieval quality. On the other hand, the ranker is trained by learning to discriminate the truth positive items from hard negative candidates sampled from the retriever. With the iteration going on, the ranker may become more precise, which in return gives rise to informative training signals for the retriever; meanwhile, with the improvement of retriever, harder negative candidates can be sampled, which contributes to a higher discriminative capability of the ranker. To facilitate the effective conduct of CoRR, an asymptotic-unbiased approximation of KL divergence is introduced for the knowledge distillation over sampled items; besides, a scalable and adaptive strategy is developed to efficiently sample from the retriever. Comprehensive experimental studies are performed over four large-scale benchmark datasets, where CoRR improves the overall recommendation quality resulting from the cooperation between retriever and ranker.
翻译:深层推荐系统共同利用检索和排序操作来生成建议结果。 检索器目标从整个项目中选择少量相关候选人, 效率高; 排序器通常比较精确, 但耗时, 通常应该以高精度来识别被检索的候选人中的最佳项目。 然而, 检索器和排序器通常以不合作的方式接受培训, 在整个工作过程中导致建议性能有限。 在这项工作中, 我们提议了一个全新的 DRS 样本培训框架 CoRR( 用于合作检索器和排序器), 使检索器和排序器能够相互加强。 一方面, 检索器是从建议数据中学习的, 并且通过知识蒸馏法, 精选器更精确, 知识蒸馏器可能会提供超弱的超视力信号, 提高检索质量。 另一方面, 排名器通过学习从回收器中抽取的硬性性级排序器, 升级器可能变得更加精确, 升级器可能从回收器和升级器之间, 升级器的回报为信息性化的升级信号, 升级后, 升级后再升级后再升为更精确, 。 升级后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后推后。 。 。 。 。 。 。 。 。 。