In this work, we study a novel class of projection-based algorithms for linearly constrained problems (LCPs) which have a lot of applications in statistics, optimization, and machine learning. Conventional primal gradient-based methods for LCPs call a projection after each (stochastic) gradient descent, resulting in that the required number of projections equals that of gradient descents (or total iterations). Motivated by the recent progress in distributed optimization, we propose the delayed projection technique that calls a projection once for a while, lowering the projection frequency and improving the projection efficiency. Accordingly, we devise a series of stochastic methods for LCPs using the technique, including a variance reduced method and an accelerated one. We theoretically show that it is feasible to improve projection efficiency in both strongly convex and generally convex cases. Our analysis is simple and unified and can be easily extended to other methods using delayed projections. When applying our new algorithms to federated optimization, a newfangled and privacy-preserving subfield in distributed optimization, we obtain not only a variance reduced federated algorithm with convergence rates better than previous works, but also the first accelerated method able to handle data heterogeneity inherent in federated optimization.
翻译:在这项工作中,我们研究了一种针对线性限制问题的新颖的基于预测的算法(LCP),这种算法在统计、优化和机器学习方面有许多应用。基于LCP的常规原始梯度方法要求在每个(随机)梯度下降后进行预测,结果所需的预测数等于梯度下降(或总迭代)数。在分配优化方面最近的进展推动下,我们提议了一种延迟的预测技术,需要对一次的预测进行一次预测,降低预测频率并改进预测效率。因此,我们设计了一系列使用该技术的LCP的随机分析方法,包括一种降低差异的方法和加速的方法。我们理论上表明,在强烈的 convex和一般的 convex情况下提高预测效率是可行的。我们的分析简单而统一,而且很容易扩大到使用延迟的预测的其他方法。在将我们的新算法应用于分流式优化、一种新缠绕和保持隐私的子领域时,我们在分配优化时,不仅获得了一种与比先前的同步率更好的压缩的节制方法,而且还获得了一种加速的方法。