Recommender systems are gaining increasing and critical impacts on human and society since a growing number of users use them for information seeking and decision making. Therefore, it is crucial to address the potential unfairness problems in recommendations. Just like users have personalized preferences on items, users' demands for fairness are also personalized in many scenarios. Therefore, it is important to provide personalized fair recommendations for users to satisfy their personalized fairness demands. Besides, previous works on fair recommendation mainly focus on association-based fairness. However, it is important to advance from associative fairness notions to causal fairness notions for assessing fairness more properly in recommender systems. Based on the above considerations, this paper focuses on achieving personalized counterfactual fairness for users in recommender systems. To this end, we introduce a framework for achieving counterfactually fair recommendations through adversary learning by generating feature-independent user embeddings for recommendation. The framework allows recommender systems to achieve personalized fairness for users while also covering non-personalized situations. Experiments on two real-world datasets with shallow and deep recommendation algorithms show that our method can generate fairer recommendations for users with a desirable recommendation performance.
翻译:由于越来越多的用户利用建议系统进行信息搜索和决策,建议系统正在对人类和社会产生越来越大的重大影响,因此,必须解决建议中潜在的不公平问题。正如用户对项目有个性化偏好一样,许多情况下用户对公平的要求也是个性化的。因此,重要的是为用户提供个性化公平建议,以满足其个性化公平要求。此外,以往关于公平建议的工作主要侧重于基于协会的公平。然而,重要的是,必须从将公平概念联系起来,转向因果关系公平概念,以便更恰当地评估建议系统中的公平。根据上述考虑,本文件侧重于在建议系统中实现用户个性化反事实公平。为此,我们引入了一个框架,通过建立依赖特性的用户嵌入式建议,通过对抗性学习实现反实际公平建议。框架允许推荐系统实现用户个性化公平,同时涵盖非个性化情况。用浅厚的建议算法对两个真实世界数据集进行实验,表明我们的方法可以为用户提出更公平的建议,建议业绩良好。