Distributed methods for training models on graph datasets have recently grown in popularity, due to the size of graph datasets as well as the private nature of graphical data like social networks. However, the graphical structure of this data means that it cannot be disjointly partitioned between different learning clients, leading to either significant communication overhead between clients or a loss of information available to the training method. We introduce Federated Graph Convolutional Network (FedGCN), which uses federated learning to train GCN models with optimized convergence rate and communication cost. Compared to prior methods that require communication among clients at each iteration, FedGCN preserves the privacy of client data and only needs communication at the initial step, which greatly reduces communication cost and speeds up the convergence rate. We theoretically analyze the tradeoff between FedGCN's convergence rate and communication cost under different data distributions, introducing a general framework can be generally used for the analysis of all edge-completion-based GCN training algorithms. Experimental results demonstrate the effectiveness of our algorithm and validate our theoretical analysis.
翻译:由于图表数据集的规模以及社交网络等图形数据的私人性质,图表数据集培训模型的分布方法最近越来越受欢迎。然而,这些数据的图形结构意味着它不能分散在不同学习客户之间,从而导致客户之间的大量通信间接费用或培训方法所能获得的信息丢失。我们引入了Freederal Place Convolutional Network(FedGCN),它利用联合学习对GCN模型进行优化趋同率和通信成本的培训。与需要客户在每次循环中进行沟通的先前方法相比,FedGCN保存客户数据的隐私,只需在初始阶段进行沟通,从而大大降低通信成本,加快聚合速度。我们从理论上分析了FedGCN在不同数据分配下合并率和通信成本之间的权衡,引入一个一般用于分析所有边缘完成GCN培训算法的一般框架。实验结果表明我们的算法的有效性,并验证我们的理论分析。