Learning large-scale pre-trained models on broad-ranging data and then transfer to a wide range of target tasks has become the de facto paradigm in many machine learning (ML) communities. Such big models are not only strong performers in practice but also offer a promising way to break out of the task-specific modeling restrictions, thereby enabling task-agnostic and unified ML systems. However, such a popular paradigm is mainly unexplored by the recommender systems (RS) community. A critical issue is that standard recommendation models are primarily built on categorical identity features. That is, the users and the interacted items are represented by their unique IDs, which are generally not shareable across different systems or platforms. To pursue the transferable recommendations, we propose studying pre-trained RS models in a novel scenario where a user's interaction feedback involves a mixture-of-modality (MoM) items, e.g., text and images. We then present TransRec, a very simple modification made on the popular ID-based RS framework. TransRec learns directly from the raw features of the MoM items in an end-to-end training manner and thus enables effective transfer learning under various scenarios without relying on overlapped users or items. We empirically study the transferring ability of TransRec across four different real-world recommendation settings. Besides, we look at its effects by scaling source and target data size. Our results suggest that learning neural recommendation models from MoM feedback provides a promising way to realize universal RS.
翻译:在许多机器学习(ML)社区中,这些大型模型不仅在实践中表现良好,而且提供了打破任务特定模型限制的有希望的方法,从而可以实现任务性模型限制,从而使任务性模型系统能够实现任务性模型和统一的ML系统。然而,这种广受欢迎的模式主要没有被推荐者系统(RS)界所探索。一个关键问题是标准建议模式主要建立在绝对的识别特征基础上。这就是,用户和互动项目由它们独特的ID代表,通常不能在不同系统或平台中分享。为了落实可转让的建议,我们提议在一种新的假设中研究预先培训的RS模式,在这种新假设中,用户的互动反馈涉及混合式(MOM)项目,例如文本和图像。我们然后介绍TransRec,在基于身份的通用框架上作出非常简单的修改。 TransRec直接从IM项目的原始特征中学习,通常不能在不同的系统或平台上分享。我们建议性指标性模型在不依赖标准性模型的情况下,可以有效地转移我们建议性的项目。