Deep learning has achieved great success in many applications. However, its deployment in practice has been hurdled by two issues: the privacy of data that has to be aggregated centrally for model training and high communication overhead due to transmission of a large amount of data usually geographically distributed. Addressing both issues is challenging and most existing works could not provide an efficient solution. In this paper, we develop FedPC, a Federated Deep Learning Framework for Privacy Preservation and Communication Efficiency. The framework allows a model to be learned on multiple private datasets while not revealing any information of training data, even with intermediate data. The framework also minimizes the amount of data exchanged to update the model. We formally prove the convergence of the learning model when training with FedPC and its privacy-preserving property. We perform extensive experiments to evaluate the performance of FedPC in terms of the approximation to the upper-bound performance (when training centrally) and communication overhead. The results show that FedPC maintains the performance approximation of the models within $8.5\%$ of the centrally-trained models when data is distributed to 10 computing nodes. FedPC also reduces the communication overhead by up to $42.20\%$ compared to existing works.
翻译:深层学习在许多应用中取得了巨大成功。然而,其实际应用却因两个问题而受到了阻碍:数据隐私,必须集中汇总,用于示范培训,以及由于传输通常地理分布的大量数据而导致的通信管理费用高昂。解决这两个问题具有挑战性,而且大多数现有工作无法提供有效的解决办法。在本文件中,我们开发了FedPC,一个隐私保护和通信效率的联邦深层学习框架。该框架允许在多个私人数据集上学习模型,同时不透露任何培训数据信息,甚至没有中间数据。框架还最大限度地减少了为更新模型而交换的数据数量。在与美联储培训及其隐私保护财产时,我们正式证明了学习模式的趋同。我们进行了广泛的实验,以评价美联储在接近(中央培训)和通信管理费用方面的业绩。结果显示,美联储在向10个计算节点分发数据时,将模型的性能近似值维持在8.5美元之内。美联储还将通信管理费用比现有工作减少42.20美元。