Communication efficiency and robustness are two major issues in modern distributed learning framework. This is due to the practical situations where some computing nodes may have limited communication power or may behave adversarial behaviors. To address the two issues simultaneously, this paper develops two communication-efficient and robust distributed learning algorithms for convex problems. Our motivation is based on surrogate likelihood framework and the median and trimmed mean operations. Particularly, the proposed algorithms are provably robust against Byzantine failures, and also achieve optimal statistical rates for strong convex losses and convex (non-smooth) penalties. For typical statistical models such as generalized linear models, our results show that statistical errors dominate optimization errors in finite iterations. Simulated and real data experiments are conducted to demonstrate the numerical performance of our algorithms.
翻译:通信效率和稳健度是现代分布式学习框架中的两个主要问题,其原因是一些计算节点的通信能力有限或可能采取对抗行为的实际情况。为了同时解决这两个问题,本文件开发了两种高效且稳健的通信分布式学习算法,用于解决二次曲线问题。我们的动机是基于代用可能性框架以及中位和缩小平均操作。特别是,拟议的算法对拜占庭失败具有可辨称的稳健性,并且对于严重的二次曲线损失和二次曲线(非模拟)处罚也达到了最佳的统计率。对于典型的统计模型,如通用线性模型,我们的结果显示统计错误主导了有限迭代中的优化错误。进行了模拟和真实的数据实验,以展示我们算法的数值性能。