In this paper, we initiate a study of functional minimization in Federated Learning. First, in the semi-heterogeneous setting, when the marginal distributions of the feature vectors on client machines are identical, we develop the federated functional gradient boosting (FFGB) method that provably converges to the global minimum. Subsequently, we extend our results to the fully-heterogeneous setting (where marginal distributions of feature vectors may differ) by designing an efficient variant of FFGB called FFGB.C, with provable convergence to a neighborhood of the global minimum within a radius that depends on the total variation distances between the client feature distributions. For the special case of square loss, but still in the fully heterogeneous setting, we design the FFGB.L method that also enjoys provable convergence to a neighborhood of the global minimum but within a radius depending on the much tighter Wasserstein-1 distances. For both FFGB.C and FFGB.L, the radii of convergence shrink to zero as the feature distributions become more homogeneous. Finally, we conduct proof-of-concept experiments to demonstrate the benefits of our approach against natural baselines.
翻译:在本文中,我们开始研究联邦学习中的功能最小化。首先,在半异质环境下,当客户机器上特性矢量的边际分布相同时,我们开发了可以与全球最小值一致的联盟性功能梯度加速法(FFGB)方法。随后,我们通过设计一个称为FFGB.C和FFGB.C的高效变体,将结果推广到全异性环境(其中特性矢量的边际分布可能不同 ), 在取决于客户特征分布之间完全差异距离的半径范围内与全球最低值的相邻相融合。对于典型损失的特殊案例,我们仍然在完全混合的环境下,我们设计了FFGB.L方法,该方法也可以与全球最小值相近,但在半径小范围内,取决于远较紧的Wasserstein-1距离。对于FFGB.C和FFGB.L来说,趋同的弧度会缩小为零,因为特征分布变得更加一致。最后,我们用证据和概念实验来证明我们自然基线方法的好处。