In order to develop effective sequential recommenders, a series of sequence representation learning (SRL) methods are proposed to model historical user behaviors. Most existing SRL methods rely on explicit item IDs for developing the sequence models to better capture user preference. Though effective to some extent, these methods are difficult to be transferred to new recommendation scenarios, due to the limitation by explicitly modeling item IDs. To tackle this issue, we present a novel universal sequence representation learning approach, named UniSRec. The proposed approach utilizes the associated description text of items to learn transferable representations across different recommendation scenarios. For learning universal item representations, we design a lightweight item encoding architecture based on parametric whitening and mixture-of-experts enhanced adaptor. For learning universal sequence representations, we introduce two contrastive pre-training tasks by sampling multi-domain negatives. With the pre-trained universal sequence representation model, our approach can be effectively transferred to new recommendation domains or platforms in a parameter-efficient way, under either inductive or transductive settings. Extensive experiments conducted on real-world datasets demonstrate the effectiveness of the proposed approach. Especially, our approach also leads to a performance improvement in a cross-platform setting, showing the strong transferability of the proposed universal SRL method. The code and pre-trained model are available at: https://github.com/RUCAIBox/UniSRec.
翻译:为了制定有效的相继建议,提议了一系列序列代表学习方法,以模拟历史用户行为;大多数现有的SRL方法依靠明确的项目编号来开发序列模型,以更好地捕捉用户偏好;虽然这些方法在某种程度上是有效的,但由于明确模拟项目ID的局限性,难以转移到新的建议设想方案;为了解决这一问题,我们提出了一个全新的通用顺序代表学习方法,名为UniSRec。拟议方法利用相关项目说明文本来学习不同建议情景的可转让表达方式。为学习通用项目表示方式,我们设计了一个以参数白化为基础的轻量项目编码结构,而专家混合调整器则得到加强。为学习通用序列表示方式,我们采用两种对比式的培训前任务,通过抽样多位负值。由于经过事先培训的普遍排序代表模式,我们的方法可以有效地以参数效率的方式转移到新的建议领域或平台,在不感动或感动的环境中学习可转移的表述。在现实世界数据集上进行的广泛实验显示了拟议通用项目表示的方法的有效性。特别是,我们的方法通过抽样评估,还导致在常规/经调整的系统前的可操作方法改进。