Recently, multi-interest models, which extract interests of a user as multiple representation vectors, have shown promising performances for sequential recommendation. However, none of existing multi-interest recommendation models consider the Out-Of-Distribution (OOD) generalization problem, in which interest distribution may change. Considering multiple interests of a user are usually highly correlated, the model has chance to learn spurious correlations between noisy interests and target items. Once the data distribution changes, the correlations among interests may also change, and the spurious correlations will mislead the model to make wrong predictions. To tackle with above OOD generalization problem, we propose a novel multi-interest network, named DEep Stable Multi-Interest Learning (DESMIL), which attempts to de-correlate the extracted interests in the model, and thus spurious correlations can be eliminated. DESMIL applies an attentive module to extract multiple interests, and then selects the most important one for making final predictions. Meanwhile, DESMIL incorporates a weighted correlation estimation loss based on Hilbert-Schmidt Independence Criterion (HSIC), with which training samples are weighted, to minimize the correlations among extracted interests. Extensive experiments have been conducted under both OOD and random settings, and up to 36.8% and 21.7% relative improvements are achieved respectively.
翻译:最近,多兴趣模型(multi-interest models),这些模型将用户的兴趣作为多个表示向量提取出来,在序列推荐中表现出了出色的性能。然而,现有的多兴趣推荐模型都没有考虑Out-Of-Distribution (OOD) 一般化问题,即兴趣分布可能会发生变化。考虑到用户的多个兴趣通常高度相关,因此模型有机会学习嘈杂兴趣和目标项目之间的虚假相关性。一旦数据分布发生变化,兴趣之间的相关性也可能发生变化, 虚假相关将会误导模型做出错误预测。为了解决上述OOD泛化问题,我们提出了一个新颖的多兴趣网络,称为DEep Stable Multi-Interest Learning (DESMIL),它试图在模型中解耦提取的兴趣,从而可以消除虚假相关。 DESMIL应用一个关注模块提取多个兴趣,然后选择最重要的兴趣作为最终预测。同时,DESMIL结合基于Hilbert-Schmidt 独立原则(HSIC)的加权相关估计损失,对训练样本进行加权,以最小化提取的兴趣之间的相关性。在OOD和随机设置下进行了广泛的实验,分别获得了36.8%和21.7%的相对改进。