News recommendation aims to display news articles to users based on their personal interest. Existing news recommendation methods rely on centralized storage of user behavior data for model training, which may lead to privacy concerns and risks due to the privacy-sensitive nature of user behaviors. In this paper, we propose a privacy-preserving method for news recommendation model training based on federated learning, where the user behavior data is locally stored on user devices. Our method can leverage the useful information in the behaviors of massive number users to train accurate news recommendation models and meanwhile remove the need of centralized storage of them. More specifically, on each user device we keep a local copy of the news recommendation model, and compute gradients of the local model based on the user behaviors in this device. The local gradients from a group of randomly selected users are uploaded to server, which are further aggregated to update the global model in the server. Since the model gradients may contain some implicit private information, we apply local differential privacy (LDP) to them before uploading for better privacy protection. The updated global model is then distributed to each user device for local model update. We repeat this process for multiple rounds. Extensive experiments on a real-world dataset show the effectiveness of our method in news recommendation model training with privacy protection.
翻译:新闻建议旨在根据用户的个人兴趣向用户展示新闻文章; 现有的新闻建议方法依靠集中储存用户行为数据,用于示范培训,这可能导致隐私问题和风险,因为用户行为具有隐私敏感性。 在本文中,我们提议以隐私保护方法为基于联合学习的新闻建议模式培训提供隐私保护方法,用户行为数据在当地存储在用户设备中,用户行为数据可以储存在用户设备中。我们的方法可以利用大量用户的行为中的有用信息来培训准确的新闻建议模式,同时消除集中储存模式的需要。更具体地说,我们保存每个用户设备的新闻建议模式的本地副本,并计算基于该设备用户行为的本地模式的梯度。随机选择的用户组的本地梯度被上传到服务器,进一步汇总以更新服务器中的全球模式。由于模型梯度可能包含一些隐含的私人信息,我们在上传更好的隐私保护之前,先对他们应用本地差异隐私。 更新后的全球模式被分发给每个用户设备,用于更新本地模式。 我们重复了这个程序, 以多轮的隐私保护方式进行真正的数据测试。