In federated learning (FL), a machine learning model is trained on multiple nodes in a decentralized manner, while keeping the data local and not shared with other nodes. However, FL requires the nodes to also send information on the model parameters to a central server for aggregation. However, the information sent from the nodes to the server may reveal some details about each node's local data, thus raising privacy concerns. Furthermore, the repetitive uplink transmission from the nodes to the server may result in a communication overhead and network congestion. To address these two challenges, in this paper, a novel two-bit aggregation algorithm is proposed with guaranteed differential privacy and reduced uplink communication overhead. Extensive experiments demonstrate that the proposed aggregation algorithm can achieve the same performance as state-of-the-art approaches on datasets such as MNIST, Fashion MNIST, CIFAR-10, and CIFAR-100, while ensuring differential privacy and improving communication efficiency.
翻译:在联合学习(FL)中,一个机器学习模式以分散方式就多个节点进行了培训,同时保留了当地数据,而没有与其他节点共享。然而,FL要求节点也向中央服务器发送模型参数信息,以便汇总;然而,节点向服务器发送的信息可能揭示每个节点当地数据的一些细节,从而引起隐私问题。此外,节点向服务器的重复连接传输可能导致通信间接费用和网络拥堵。为了应对这两个挑战,本文件提出了一个新的双位组合算法,其中保证有不同的隐私和减少连接通信间接费用。广泛的实验表明,拟议的汇总算法可以取得与诸如MNIST、Fashon MNIST、CIFAR-10和CIFAR-100等最新数据集方法相同的性能,同时确保不同的隐私,提高通信效率。