We consider the problem of online stochastic optimization in a distributed setting with $M$ clients connected through a central server. We develop a distributed online learning algorithm that achieves order-optimal cumulative regret with low communication cost measured in the total number of bits transmitted over the entire learning horizon. This is in contrast to existing studies which focus on the offline measure of simple regret for learning efficiency. The holistic measure for communication cost also departs from the prevailing approach that \emph{separately} tackles the communication frequency and the number of bits in each communication round.
翻译:我们考虑在分布式环境中,通过一个中央服务器连接到$M美元客户的分布式环境中实现在线优化的问题。我们开发了一种分布式在线学习算法,以整个学习视野传输的总比特的低通信成本来衡量,从而实现秩序最佳累积遗憾,而通信成本低。这与现有研究形成对照,现有研究侧重于简单遗憾的离线度量,以便提高学习效率。通信成本的整体度量也偏离了目前采用的处理通信频率和每轮通信比特数的通用方法。