Sequential recommendation models are primarily optimized to distinguish positive samples from negative ones during training in which negative sampling serves as an essential component in learning the evolving user preferences through historical records. Except for randomly sampling negative samples from a uniformly distributed subset, many delicate methods have been proposed to mine negative samples with high quality. However, due to the inherent randomness of negative sampling, false negative samples are inevitably collected in model training. Current strategies mainly focus on removing such false negative samples, which leads to overlooking potential user interests, lack of recommendation diversity, less model robustness, and suffering from exposure bias. To this end, we propose a novel method that can Utilize False Negative samples for sequential Recommendation (UFNRec) to improve model performance. We first devise a simple strategy to extract false negative samples and then transfer these samples to positive samples in the following training process. Furthermore, we construct a teacher model to provide soft labels for false negative samples and design a consistency loss to regularize the predictions of these samples from the student model and the teacher model. To the best of our knowledge, this is the first work to utilize false negative samples instead of simply removing them for the sequential recommendation. Experiments on three benchmark public datasets are conducted using three widely applied SOTA models. The experiment results demonstrate that our proposed UFNRec can effectively draw information from false negative samples and further improve the performance of SOTA models. The code is available at https://github.com/UFNRec-code/UFNRec.
翻译:序列建议模式主要优化,在培训期间将正面样本与负面样本区分开来,在培训期间,负面抽样是通过历史记录了解用户对不断演变的偏好的基本组成部分。除了从统一分布的子集随机抽样对负面样本进行抽样抽样抽样抽样抽样抽样抽样外,还提出了许多精细的方法,但由于负面抽样固有的随机性,不可避免地在模式培训中收集虚假负面样本。目前的战略主要侧重于消除这些虚假的负面样本,这导致忽略潜在用户的利益,缺乏建议多样性,不那么模型的稳健性,并受到接触偏差的影响。为此,我们提出了一种新颖的方法,可以利用假否定的样本来提出顺序建议(UFNRec),以改进模型的性能。我们首先设计了一个简单的战略,提取虚假的负面样本,然后在接下来的培训过程中将这些样本转移到正面样本。此外,我们还建立一个教师模型,为虚假的负面样本提供软标签,并设计一致性损失,以便从学生模型和教师模型中规范对这些样本的预测。我们最了解的是,这是首先使用虚假的负面样本,而不是简单地在顺序建议中使用这些样本(UFRC/Rec),然后进行实验,然后用三个实验,然后将三个数据结果作为基准,然后进行。我们提出的三个试验。我们提出的数据标准是用来检验。我们提出的三个数据标准是用来检验。