Many modern sequential recommender systems use deep neural networks, which can effectively estimate the relevance of items but require a lot of time to train. Slow training increases expenses, hinders product development timescales and prevents the model from being regularly updated to adapt to changing user preferences. Training such sequential models involves appropriately sampling past user interactions to create a realistic training objective. The existing training objectives have limitations. For instance, next item prediction never uses the beginning of the sequence as a learning target, thereby potentially discarding valuable data. On the other hand, the item masking used by BERT4Rec is only weakly related to the goal of the sequential recommendation; therefore, it requires much more time to obtain an effective model. Hence, we propose a novel Recency-based Sampling of Sequences training objective that addresses both limitations. We apply our method to various recent and state-of-the-art model architectures - such as GRU4Rec, Caser, and SASRec. We show that the models enhanced with our method can achieve performances exceeding or very close to stateof-the-art BERT4Rec, but with much less training time.
翻译:许多现代相继建议系统使用深层神经网络,这些网络可以有效估计物品的相关性,但需要大量时间进行培训。培训缓慢增加费用,阻碍产品开发时间尺度,妨碍对模型进行定期更新以适应用户的偏好变化。这类相继模式的培训涉及对过去用户的互动进行适当抽样,以创造现实的培训目标。现有培训目标有局限性。例如,下一项项目预测从不将序列的起点作为学习目标,从而有可能丢弃有价值的数据。另一方面,BERT4Rec所使用的遮盖物与顺序建议的目标关系不大;因此,它需要更多的时间才能获得有效的模型。因此,我们提议采用基于序列的新的序列抽样抽样,以克服这两个限制。我们采用的方法适用于各种最新和最新最先进的模型结构,如GRU4Rec、Cacer和SASRec。我们表明,用我们的方法强化的模型能够达到超过或非常接近尖端的BERT4Rec的性能,但培训时间要少得多。