In many distributed learning setups such as federated learning (FL), client nodes at the edge use individually collected data to compute local gradients and send them to a central master server. The master server then aggregates the received gradients and broadcasts the aggregation to all clients, with which the clients can update the global model. In this paper, we consider multi-server federated learning with secure aggregation and unreliable communication links. We first define a threat model using Shannon's information-theoretic security framework and propose a novel scheme called Lagrange Coding with Mask (LCM), which divides the servers into groups and uses Coding and Masking techniques. LCM can achieve a trade-off between the uplink and downlink communication loads by adjusting the number of servers in each group. Furthermore, we derive the lower bounds of the uplink and downlink communication loads, respectively, and prove that LCM achieves the optimal uplink communication load, which is unrelated to the number of collusion clients.
翻译:在许多分布式学习设置中,如联邦学习 (FL),边缘端的客户端节点使用单独收集的数据计算本地梯度并将其发送到中央主服务器。主服务器然后聚合接收到的梯度并将聚合广播到所有客户端,客户端可以使用此更新全局模型。在本文中,我们考虑具有安全聚合和不可靠通信链路的多服务器联合学习。我们首先使用Shannon的信息论安全框架定义了一个威胁模型,并提出了一个称为Lagrange编码与掩码的新方案(LCM),将服务器划分为组,并使用编码和掩码技术。 LCM可以通过调整每组中的服务器数量,在上行和下行通信负载之间实现权衡。此外,我们分别导出了上行和下行通信负载的下限,并证明了LCM实现了最优上行通信负载,这与串通客户的数量无关。