In federated learning, participating clients typically possess non-i.i.d. data, posing a significant challenge to generalization to unseen distributions. To address this, we propose a Wasserstein distributionally robust optimization scheme called WAFL. Leveraging its duality, we frame WAFL as an empirical surrogate risk minimization problem, and solve it using a local SGD-based algorithm with convergence guarantees. We show that the robustness of WAFL is more general than related approaches, and the generalization bound is robust to all adversarial distributions inside the Wasserstein ball (ambiguity set). Since the center location and radius of the Wasserstein ball can be suitably modified, WAFL shows its applicability not only in robustness but also in domain adaptation. Through empirical evaluation, we demonstrate that WAFL generalizes better than the vanilla FedAvg in non-i.i.d. settings, and is more robust than other related methods in distribution shift settings. Further, using benchmark datasets we show that WAFL is capable of generalizing to unseen target domains.
翻译:在联合学习中,参与客户通常拥有非i.id.d.数据,这对推广到无形分布构成重大挑战。为了解决这个问题,我们提议一个瓦塞斯坦分配上强大的优化计划,称为WAFL。利用它的双重性,我们将WAFL作为经验性替代风险最小化问题,并使用基于本地SGD的算法和趋同保证加以解决。我们表明WAFL的稳健性比相关方法更为普遍,而通用约束对瓦塞斯坦球(雄性组合)内的所有对抗性分布都十分有力。由于瓦塞斯坦球的中心位置和半径可以适当修改,WAFL不仅在稳健性方面,而且在领域适应方面显示出其适用性。通过经验评估,我们证明WAFL在非i.d.环境中比Vanilla FDAvg普遍化,在分销转移环境中比其他相关方法更加有力。此外,我们使用基准数据集表明WAFFL能够向看不见的目标区域普及。