Collecting and training over sensitive personal data raise severe privacy concerns in personalized recommendation systems, and federated learning can potentially alleviate the problem by training models over decentralized user data.However, a theoretically private solution in both the training and serving stages of federated recommendation is essential but still lacking.Furthermore, naively applying differential privacy (DP) to the two stages in federated recommendation would fail to achieve a satisfactory trade-off between privacy and utility due to the high-dimensional characteristics of model gradients and hidden representations.In this work, we propose a federated news recommendation method for achieving a better utility in model training and online serving under a DP guarantee.We first clarify the DP definition over behavior data for each round in the life-circle of federated recommendation systems.Next, we propose a privacy-preserving online serving mechanism under this definition based on the idea of decomposing user embeddings with public basic vectors and perturbing the lower-dimensional combination coefficients. We apply a random behavior padding mechanism to reduce the required noise intensity for better utility. Besides, we design a federated recommendation model training method, which can generate effective and public basic vectors for serving while providing DP for training participants. We avoid the dimension-dependent noise for large models via label permutation and differentially private attention modules. Experiments on real-world news recommendation datasets validate that our method achieves superior utility under a DP guarantee in both training and serving of federated news recommendations.
翻译:个人敏感数据的收集和培训在个人化建议系统中引起对隐私的严重关切,而联合学习则有可能通过使用分散用户数据的培训模型来缓解这一问题。 然而,在联合建议的培训和服务阶段,一个理论上私人的解决办法是必要的,但是仍然缺乏。 此外,由于模型梯度和隐蔽的表达方式具有高度的特征,因此在隐私和实用性之间不能实现令人满意的权衡。 在这项工作中,我们提出了一个联合式的新闻建议方法,以便在模型培训中和在DP保证下在线服务中实现更好的效用。 我们首先澄清DP的定义,即在联合建议系统的生命圈中,对每一轮进行的行为数据。