Federated learning is a useful framework for centralized learning from distributed data under practical considerations of heterogeneity, asynchrony, and privacy. Federated architectures are frequently deployed in deep learning settings, which generally give rise to non-convex optimization problems. Nevertheless, most existing analysis are either limited to convex loss functions, or only establish first-order stationarity, despite the fact that saddle-points, which are first-order stationary, are known to pose bottlenecks in deep learning. We draw on recent results on the second-order optimality of stochastic gradient algorithms in centralized and decentralized settings, and establish second-order guarantees for a class of federated learning algorithms.
翻译:联邦学习是一个有用的框架,用于在对异质性、无同步性和隐私的实际考虑下,从分布的数据中集中学习。联邦建筑经常被部署在深层学习环境中,这通常会导致非凝聚优化问题。然而,大多数现有分析要么局限于锥形损失功能,要么仅仅建立一级固定状态,尽管已知第一阶固定状态的马鞍点在深层学习中构成瓶颈。我们借鉴了在集中和分散环境中第二阶的随机梯度算法的最佳性的最新结果,并为一类联邦化学习算法建立了第二阶级保障。