Despite the prominence of neural network approaches in the field of recommender systems, simple methods such as matrix factorization with quadratic loss are still used in industry for several reasons. These models can be trained with alternating least squares, which makes them easy to implement in a massively parallel manner, thus making it possible to utilize billions of events from real-world datasets. Large-scale recommender systems need to account for severe popularity skew in the distributions of users and items, so a lot of research is focused on implementing sparse, mixed dimension or shared embeddings to reduce both the number of parameters and overfitting on rare users and items. In this paper we propose two matrix factorization models with mixed dimension embeddings, which can be optimized in a massively parallel fashion using the alternating least squares approach.
翻译:尽管神经网络方法在推荐者系统领域占有突出地位,但工业中仍然出于若干原因使用简单的方法,如带有二次损失的矩阵化因子化等。这些模型可以使用交替最少的方形进行培训,这样就容易以大规模平行的方式实施,从而有可能利用来自真实世界数据集的数十亿个事件。 大型建议系统需要考虑到用户和项目分布中的严重受欢迎性偏差,因此许多研究侧重于执行稀有、混合的维度或共享嵌入,以减少参数数量,并过度适应稀有用户和项目。在本文件中,我们提议采用交替最小方形方法,以大规模平行的方式优化两种具有混合维度的矩阵化模型。