To prevent implicit privacy disclosure in sharing gradients among data owners (DOs) under federated learning (FL), differential privacy (DP) and its variants have become a common practice to offer formal privacy guarantees with low overheads. However, individual DOs generally tend to inject larger DP noises for stronger privacy provisions (which entails severe degradation of model utility), while the curator (i.e., aggregation server) aims to minimize the overall effect of added random noises for satisfactory model performance. To address this conflicting goal, we propose a novel dynamic privacy pricing (DyPP) game which allows DOs to sell individual privacy (by lowering the scale of locally added DP noise) for differentiated economic compensations (offered by the curator), thereby enhancing FL model utility. Considering multi-dimensional information asymmetry among players (e.g., DO's data distribution and privacy preference, and curator's maximum affordable payment) as well as their varying private information in distinct FL tasks, it is hard to directly attain the Nash equilibrium of the mixed-strategy DyPP game. Alternatively, we devise a fast reinforcement learning algorithm with two layers to quickly learn the optimal mixed noise-saving strategy of DOs and the optimal mixed pricing strategy of the curator without prior knowledge of players' private information. Experiments on real datasets validate the feasibility and effectiveness of the proposed scheme in terms of faster convergence speed and enhanced FL model utility with lower payment costs.
翻译:为防止在根据联合学习(FL),数据所有人之间分享梯度时隐含隐私披露,在分享数据所有人(DOs)之间分享梯度,差异隐私(DP)及其变式已成为一种常见做法,为低管理费提供正式的隐私保障;然而,个别DO通常会为更强有力的隐私规定注入更大的DP噪音(这导致模型功能严重退化),而馆长(即聚合服务器)的目的是尽量减少增加随机噪音对令人满意的模型性能的总体影响。为了实现这一相互矛盾的目标,我们提议采用新的动态隐私定价游戏(DYPP)游戏,让DO能够出售个人隐私(通过降低当地添加的DP噪音的规模),以提供差别经济补偿(由馆长提供),从而增强FL模式的效用。考虑到玩家(例如DO的数据分布和隐私偏好,以及馆长最大可负担的付款)之间的多方面信息不对称,以及他们在不同的FL任务中的不同私人信息,很难直接达到混合战略的纳什平衡。 或者,我们设计一个快速强化的通用性计算方法,在两个层次上迅速强化使用FL支付战略的精度,同时学习最优的硬性计算方法,学习硬性战略,不改进了DOPPS的硬性战略,不改进的硬化成本。