We consider distributed online learning for joint regret with communication constraints. In this setting, there are multiple agents that are connected in a graph. Each round, an adversary first activates one of the agents to issue a prediction and provides a corresponding gradient, and then the agents are allowed to send a $b$-bit message to their neighbors in the graph. All agents cooperate to control the joint regret, which is the sum of the losses of the activated agents minus the losses evaluated at the best fixed common comparator parameters $u$. We observe that it is suboptimal for agents to wait for gradients that take too long to arrive. Instead, the graph should be partitioned into local clusters that communicate among themselves. Our main result is a new method that can adapt to the optimal graph partition for the adversarial activations and gradients, where the graph partition is selected from a set of candidate partitions. A crucial building block along the way is a new algorithm for online convex optimization with delayed gradient information that is comparator-adaptive, meaning that its joint regret scales with the norm of the comparator $||u||$. We further provide near-optimal gradient compression schemes depending on the ratio of $b$ and the dimension times the diameter of the graph.
翻译:我们考虑在网上进行共享学习, 以在通信限制下共同感到遗憾。 在此设置中, 有多个代理商连接到一个图表中。 每个回合, 对手首先激活一个代理商发布预测, 并提供一个相应的梯度, 然后代理商就可以在图形中向邻居发送一美元比特信息。 所有代理商合作控制联合遗憾, 即激活代理商损失减去以最佳固定共同参照基准参数评估的损失的总和 $ 美元。 我们观察到, 代理商等待过长的梯度是次最佳的。 相反, 图表应该被分割成彼此沟通的本地组群。 我们的主要结果是一种新的方法, 可以适应用于对抗性激活和梯度的最佳图形分区。 图形分区是从一组候选分区中选择的。 沿路的一个关键建筑块是使用延迟的梯度信息进行在线连接优化的新算法, 后者是比较性精确度, 意思是其与参照者标准 $u ⁇ $ $ $ 美元 之间的联合遗憾度尺度。 我们进一步提供接近- opimimmeal graimal gramachegraphy pal im pre by by restium by by expeal by by by by by expealblegleglegleglegal