Learning on Graphs (LoG) is widely used in multi-client systems when each client has insufficient local data, and multiple clients have to share their raw data to learn a model of good quality. One scenario is to recommend items to clients with limited historical data and sharing similar preferences with other clients in a social network. On the other hand, due to the increasing demands for the protection of clients' data privacy, Federated Learning (FL) has been widely adopted: FL requires models to be trained in a multi-client system and restricts sharing of raw data among clients. The underlying potential data-sharing conflict between LoG and FL is under-explored and how to benefit from both sides is a promising problem. In this work, we first formulate the Graph Federated Learning (GFL) problem that unifies LoG and FL in multi-client systems and then propose sharing hidden representation instead of the raw data of neighbors to protect data privacy as a solution. To overcome the biased gradient problem in GFL, we provide a gradient estimation method and its convergence analysis under the non-convex objective. In experiments, we evaluate our method in classification tasks on graphs. Our experiment shows a good match between our theory and the practice.
翻译:在多客户系统中,当每个客户都没有足够的本地数据,而且多个客户必须分享原始数据以学习高质量的模型时,图上学习被广泛用于多客户系统。一种设想是,向历史数据有限的客户推荐项目,并与社会网络中的其他客户分享类似的偏好。另一方面,由于对保护客户数据隐私的需求不断增加,联邦学习(FL)被广泛采用:FL要求模型在多客户系统中接受培训,并限制客户之间分享原始数据。LoG和FL之间潜在的数据共享冲突正在调查中,如何从双方受益是一个大有希望的问题。在这项工作中,我们首先将项目推荐给历史数据有限的客户,并与社会网络中的其他客户分享类似的偏好。另一方面,由于对保护客户数据隐私的需求不断增加,Falde Lead(FL)被广泛采用:FLL要求模型在多客户系统中接受培训,并限制客户之间分享原始数据。为了克服GLF的偏差问题,我们提供了一种梯度估计方法,并在非convex目标下进行其趋同分析。在实验中,我们评估了图表理论和图表理论上的方法。我们的实验显示一种良好的匹配。我们的做法。