This paper studies the problem of federated learning (FL) in the absence of a trustworthy server/clients. In this setting, each client needs to ensure the privacy of its own data without relying on the server or other clients. We study local differential privacy (LDP) and provide tight upper and lower bounds that establish the minimax optimal rates (up to logarithms) for LDP convex/strongly convex federated stochastic optimization. Our rates match the optimal statistical rates in certain practical parameter regimes ("privacy for free"). Second, we develop a novel time-varying noisy SGD algorithm, leading to the first non-trivial LDP risk bounds for FL with non-i.i.d. clients. Third, we consider the special case where each client's loss function is empirical and develop an accelerated LDP FL algorithm to improve communication complexity compared to existing works. We also provide matching lower bounds, establishing the optimality of our algorithm for convex/strongly convex settings. Fourth, with a secure shuffler to anonymize client reports (but without a trusted server), our algorithm attains the optimal central DP rates for stochastic convex/strongly convex optimization, thereby achieving optimality in the local and central models simultaneously. Our upper bounds quantify the role of network communication reliability in performance.
翻译:本文研究在缺少可靠的服务器/客户端情况下的联结学习( FL) 问题。 在此环境下, 每个客户都需要确保自己数据的隐私, 而不依赖于服务器或其他客户。 我们研究本地差异隐私( LDP), 并提供紧紧的上下下界, 以建立本地DP convex/ 强调的联结同步优化的最小最大最佳比率( 直至对数) 。 我们的费率符合某些实用参数系统中的最佳统计率 ( “ 免费隐私” ) 。 第二, 我们开发了一个新的时间变化的噪音 SGD 算法, 导致首次非三重 LDP 与非i. id. 客户的 LDP 风险约束。 第三, 我们考虑每个客户的损失功能都是经验性的特例, 并开发一个加速的 LDP FL 算法, 以提高与现有工程相比的通信复杂性。 我们还提供匹配的下限, 建立我们配置/ 组合设置的最优化的算法 。 第四, 安全地平移到本地客户端平调化的客户端平整的客户端端端端端平整的, 实现我们最优化的客户端端平整的客户端平整的平整的平整的平整的平整的平整的平整的平整的平整的平整的平局。