The issue of potential privacy leakage during centralized AI's model training has drawn intensive concern from the public. A Parallel and Distributed Computing (or PDC) scheme, termed Federated Learning (FL), has emerged as a new paradigm to cope with the privacy issue by allowing clients to perform model training locally, without the necessity to upload their personal sensitive data. In FL, the number of clients could be sufficiently large, but the bandwidth available for model distribution and re-upload is quite limited, making it sensible to only involve part of the volunteers to participate in the training process. The client selection policy is critical to an FL process in terms of training efficiency, the final model's quality as well as fairness. In this paper, we will model the fairness guaranteed client selection as a Lyapunov optimization problem and then a C2MAB-based method is proposed for estimation of the model exchange time between each client and the server, based on which we design a fairness guaranteed algorithm termed RBCS-F for problem-solving. The regret of RBCS-F is strictly bounded by a finite constant, justifying its theoretical feasibility. Barring the theoretical results, more empirical data can be derived from our real training experiments on public datasets.
翻译:在集中的AI模式培训中,潜在的隐私渗漏问题引起了公众的高度关注。一个称为Federed Learning(FL)的平行和分布式计算机(或PDC)计划(或PDC)计划(即Federal Learning(FL))已成为处理隐私问题的新范例,它允许客户在当地进行示范培训,而不必上传个人敏感数据。在FL中,客户数量可能足够大,但用于模式分发和再加载的带宽非常有限,因此只有部分志愿人员参与培训过程才明智。客户选择政策对于FL进程在培训效率、最后模式质量以及公平方面至关重要。在本文件中,我们将将公平性保证客户选择作为Lyapunov优化问题,然后提出基于C2MAB的方法,用于估算每个客户与服务器之间的模式交换时间,我们据此设计一种公平保证的算法,称为RBCS-F,用于解决问题。RBCS-F的遗憾是有限的固定不变的,说明其理论可行性。我们从实际培训中可以得出更多的实验数据。