Federated learning allows mobile clients to jointly train a global model without sending their private data to a central server. Extensive works have studied the performance guarantee of the global model, however, it is still unclear how each individual client influences the collaborative training process. In this work, we defined a new notion, called {\em Fed-Influence}, to quantify this influence over the model parameters, and proposed an effective and efficient algorithm to estimate this metric. In particular, our design satisfies several desirable properties: (1) it requires neither retraining nor retracing, adding only linear computational overhead to clients and the server; (2) it strictly maintains the tenets of federated learning, without revealing any client's local private data; and (3) it works well on both convex and non-convex loss functions, and does not require the final model to be optimal. Empirical results on a synthetic dataset and the FEMNIST dataset demonstrate that our estimation method can approximate Fed-Influence with small bias. Further, we show an application of Fed-Influence in model debugging.
翻译:联邦学习使流动客户可以在不将其私人数据发送到中央服务器的情况下联合培训一个全球模型。广泛的工程研究了全球模型的绩效保障,然而,仍然不清楚每个客户如何影响协作培训过程。在这项工作中,我们界定了一个新的概念,称为`em Fed-Iffect',以量化这种对模型参数的影响,并提议了一种有效和高效的算法来估计这一指标。特别是,我们的设计满足了若干可取的特性:(1) 它既不需要再培训,也不需要收回,只给客户和服务器增加线性计算间接费用;(2) 它严格维护联合学习原则,不披露任何客户的本地私人数据;(3) 它在 convex 和非 convex 损失功能上运作良好,不需要最终模型是最佳的。合成数据集和FEMNIST数据集的实证结果表明,我们的估算方法可以以小的偏差接近Fed-Ipt。此外,我们展示了在模型调控中应用Fed-Id-Ipt。