Federated learning allows a group of distributed clients to train a common machine learning model on private data. The exchange of model updates is managed either by a central entity or in a decentralized way, e.g. by a blockchain. However, the strong generalization across all clients makes these approaches unsuited for non-independent and identically distributed (non-IID) data. We propose a unified approach to decentralization and personalization in federated learning that is based on a directed acyclic graph (DAG) of model updates. Instead of training a single global model, clients specialize on their local data while using the model updates from other clients dependent on the similarity of their respective data. This specialization implicitly emerges from the DAG-based communication and selection of model updates. Thus, we enable the evolution of specialized models, which focus on a subset of the data and therefore cover non-IID data better than federated learning in a centralized or blockchain-based setup. To the best of our knowledge, the proposed solution is the first to unite personalization and poisoning robustness in fully decentralized federated learning. Our evaluation shows that the specialization of models emerges directly from the DAG-based communication of model updates on three different datasets. Furthermore, we show stable model accuracy and less variance across clients when compared to federated averaging.
翻译:联邦学习使一组分布式客户能够就私人数据培训一个共同的机器学习模式。模式更新的交流要么由一个中央实体管理,要么以分散式方式管理,例如由一个链条管理。然而,对所有客户的有力普及使得这些方法不适合非独立和相同分布(非IID)数据。我们建议采用统一的方法,在基于定向循环图的模型更新基础上,在联合学习中进行权力下放和个人化。除了培训单一的全球模式外,客户在使用取决于各自数据相似性的其他客户的模型更新时,专门使用其本地数据。这种专业化隐含地产生于基于DAG的通信和选择模式更新。因此,我们促成专门模型的演变,侧重于一组数据,从而涵盖非IID数据,比集中式或基于链式更新的混合式学习(DAG)更好地覆盖非II数据。据我们所知,拟议的解决办法是首先将个人化统一起来,在完全分散化的模型学习中使地方数据更加稳健。我们的评价显示,在基于不同格式的模型更新时,将三个标准化的分类化数据直接显示我们基于不同格式的标准化。