Graph neural networks (GNNs) have shown their superiority in modeling graph data. Owing to the advantages of federated learning, federated graph learning (FGL) enables clients to train strong GNN models in a distributed manner without sharing their private data. A core challenge in federated systems is the non-IID problem, which also widely exists in real-world graph data. For example, local data of clients may come from diverse datasets or even domains, e.g., social networks and molecules, increasing the difficulty for FGL methods to capture commonly shared knowledge and learn a generalized encoder. From real-world graph datasets, we observe that some structural properties are shared by various domains, presenting great potential for sharing structural knowledge in FGL. Inspired by this, we propose FedStar, an FGL framework that extracts and shares the common underlying structure information for inter-graph federated learning tasks. To explicitly extract the structure information rather than encoding them along with the node features, we define structure embeddings and encode them with an independent structure encoder. Then, the structure encoder is shared across clients while the feature-based knowledge is learned in a personalized way, making FedStar capable of capturing more structure-based domain-invariant information and avoiding feature misalignment issues. We perform extensive experiments over both cross-dataset and cross-domain non-IID FGL settings, demonstrating the superiority of FedStar.
翻译:由于联合学习的好处,联合图形学习(FGL)使客户能够以分布式的方式培训强大的GNN模型,而不分享其私人数据。联邦系统的一个核心挑战是非IID问题,这个问题在现实世界图形数据中也广泛存在。例如,客户的本地数据可能来自不同的数据集,甚至领域,例如社交网络和分子,这增加了FGL获取共同共享知识和学习通用编码器的难度。从真实世界图形数据集中,我们看到一些结构属性被不同领域共享,这为分享FGL的结构知识带来巨大潜力。受此启发,我们提议FDStar,这是一个FGL框架,提取并分享用于跨世界图像化学习任务的共同基本结构信息。为了明确提取结构信息,而不是将信息与节点特性一起编码,我们定义了结构嵌入并用一个独立的结构编码。然后,在基于真实世界图形的图表设置中,能够显示FDStar的系统域域域图的跨主题结构是共享的,同时在基于我们基于常规的FDSilveral化的跨域域图中,同时学习了FStar的跨域域图。