Federated Learning(FL) is a distributed machine learning paradigm where data is distributed among clients who collaboratively train a model in a computation process coordinated by a central server. By assigning a weight to each client based on the proportion of data instances it possesses, the rate of convergence to an accurate joint model can be greatly accelerated. Some previous works studied FLin a Byzantine setting, in which a fraction of the clients may send arbitrary or even malicious information regarding their model. However, these works either ignore the issue of data unbalancedness altogether or assume that client weights are apriori known to the server, whereas, in practice, it is likely that weights will be reported to the server by the clients themselves and therefore cannot be relied upon. We address this issue for the first time by proposing a practical weight-truncation-based preprocessing method and demonstrating empirically that it is able to strike a good balance between model quality and Byzantine robustness. We also establish analytically that our method can be applied to a randomly selected sample of client weights.
翻译:联邦学习组织(FL)是一个分布式的机器学习模式,在中央服务器协调的计算过程中向合作培训模型的客户分发数据。通过根据每个客户掌握的数据比例给每个客户分配权重,可以大大加快与准确的联合模型的趋同速度。以前的一些著作研究了拜占庭环境的FLin,其中一部分客户可以随意发送有关其模型的信息,甚至是恶意信息。然而,这些作品要么忽视了数据全面不平衡的问题,要么假定服务器最了解客户的重量,而在实践中,可能客户自己会向服务器报告重量,因此无法依赖。我们第一次通过提出实用的重力调整前处理方法来解决这一问题,并用经验证明它能够在模型质量和Byzantine稳健之间取得良好的平衡。我们还通过分析确定,我们的方法可以适用于随机选择的客户重量样本。