Many important application domains generate distributed collections of heterogeneous local datasets. These local datasets are often related via an intrinsic network structure that arises from domain-specific notions of similarity between local datasets. Different notions of similarity are induced by spatiotemporal proximity, statistical dependencies, or functional relations. We use this network structure to adaptively pool similar local datasets into nearly homogenous training sets for learning tailored models. Our main conceptual contribution is to formulate networked federated learning using the concept of generalized total variation (GTV) minimization as a regularizer. This formulation is highly flexible and can be combined with almost any parametric model including Lasso or deep neural networks. We unify and considerably extend some well-known approaches to federated multi-task learning. Our main algorithmic contribution is a novel federated learning algorithm that is well suited for distributed computing environments such as edge computing over wireless networks. This algorithm is robust against model misspecification and numerical errors arising from limited computational resources including processing time or wireless channel bandwidth. As our main technical contribution, we offer precise conditions on the local models as well on their network structure such that our algorithm learns nearly optimal local models. Our analysis reveals an interesting interplay between the (information-) geometry of local models and the (cluster-) geometry of their network.
翻译:许多重要的应用领域产生分散的本地数据集。这些本地数据集往往通过一个内在的网络结构,由特定领域的地方数据集的相似概念产生。不同相似的概念是由时空相近、统计依赖性或功能关系引起的。我们使用这个网络结构,将类似的本地数据集适应性地汇集到几乎同质的学习定制模型中。我们的主要概念贡献是利用普遍的全面变异(GTV)最小化作为常规化工具的概念来建立网络化的联合会学习。这一表述非常灵活,可以与几乎任何准数模型(包括Lasso或深神经网络)相结合。我们统一并大大扩展一些众所周知的多任务联合学习方法。我们的主要算法贡献是一种新型的联邦化学习算法,它适合于分布式的计算环境,例如边际计算而不是无线网络。我们的主要概念贡献是利用有限的计算资源(包括处理时间或无线频道带宽带带带带带带宽)来建立网络化的网络化和数字错误。我们的主要技术贡献是,我们为本地模型提供了精确的条件,作为网络结构模型之间最精确的地理测量模型(我们最优化的模型)分析。