Historical user-item interaction datasets are essential in training modern recommender systems for predicting user preferences. However, the arbitrary user behaviors in most recommendation scenarios lead to a large volume of noisy data instances being recorded, which cannot fully represent their true interests. While a large number of denoising studies are emerging in the recommender system community, all of them suffer from highly dynamic data distributions. In this paper, we propose a Deep Reinforcement Learning (DRL) based framework, AutoDenoise, with an Instance Denoising Policy Network, for denoising data instances with an instance selection manner in deep recommender systems. To be specific, AutoDenoise serves as an agent in DRL to adaptively select noise-free and predictive data instances, which can then be utilized directly in training representative recommendation models. In addition, we design an alternate two-phase optimization strategy to train and validate the AutoDenoise properly. In the searching phase, we aim to train the policy network with the capacity of instance denoising; in the validation phase, we find out and evaluate the denoised subset of data instances selected by the trained policy network, so as to validate its denoising ability. We conduct extensive experiments to validate the effectiveness of AutoDenoise combined with multiple representative recommender system models.
翻译:历史用户-项目互动数据集对于培训预测用户偏好的现代建议系统至关重要。然而,大多数建议情景中的任意用户行为导致大量繁琐的数据案例被记录,无法充分代表他们的真正利益。虽然在推荐者系统社区正在出现大量分解研究,但所有这些研究都存在高度动态的数据分布。在本文件中,我们提出了一个基于深强化学习(DRL)的框架AutoDenoise(AutoDenoise),并有一个“不记名政策网络”,用于在深推荐者系统中以实例选择方式对数据案例进行分解。具体地说,AutoDenoise是DRL中适应性地选择无噪音和预测性数据案例的代理,然后直接用于培训代议人建议模型。此外,我们设计了一个两阶段的替代优化战略,以适当培训和验证“AutoDenoise”(Doise)系统。在搜索阶段,我们的目标是以实例分解能力对政策网络进行培训;在验证阶段,我们发现和评估经过培训的政策网络所选择的数据案例的分解的分解分解部分。我们发现和评估了经过培训的数据案例的分解的分解的分解分解部分。我们建议,以广泛地验证系统进行多种验证能力,以验证系统进行自我验证。我们建议进行自我验证。</s>