Multi-task learning has been widely used in real-world recommenders to predict different types of user feedback. Most prior works focus on designing network architectures for bottom layers as a means to share the knowledge about input features representations. However, since they adopt task-specific binary labels as supervised signals for training, the knowledge about how to accurately rank items is not fully shared across tasks. In this paper, we aim to enhance knowledge transfer for multi-task personalized recommendat optimization objectives. We propose a Cross-Task Knowledge Distillation (CrossDistil) framework in recommendation, which consists of three procedures. 1) Task Augmentation: We introduce auxiliary tasks with quadruplet loss functions to capture cross-task fine-grained ranking information, which could avoid task conflicts by preserving the cross-task consistent knowledge; 2) Knowledge Distillation: We design a knowledge distillation approach based on augmented tasks for sharing ranking knowledge, where tasks' predictions are aligned with a calibration process; 3) Model Training: Teacher and student models are trained in an end-to-end manner, with a novel error correction mechanism to speed up model training and improve knowledge quality. Comprehensive experiments on a public dataset and our production dataset are carried out to verify the effectiveness of CrossDistil as well as the necessity of its key components.
翻译:多任务学习被广泛用于现实世界建议者,以预测不同类型的用户反馈。大多数先前的工作重点是设计底层层网络架构,以分享关于投入特征表示的知识。然而,由于它们采用特定任务二进制标签作为监督的培训信号,如何准确排列项目的知识没有在任务之间充分共享。在本文件中,我们的目标是加强知识转让,以实现多任务个性化的建议优化目标。我们建议采用跨任务知识蒸馏(CrossDistilt)框架,由三个程序组成。 1 任务增加:我们引入带有四轮式损失功能的辅助任务,以获取跨任务细细细分级信息,这可以通过保持交叉任务的一致性知识避免任务冲突。 2 知识蒸馏:我们设计一种知识蒸馏方法,其基础是扩大知识共享的任务,任务预测与校准进程相一致;3 模式:师生模型以端到端方式接受培训,同时引入新的错误纠正机制,以加快模型培训速度,改进数据质量,作为公共数据有效性的测试。