In multi-party collaborative learning, the parameter server sends a global model to each data holder for local training and then aggregates committed models globally to achieve privacy protection. However, both the dragger issue of synchronous collaborative learning and the staleness issue of asynchronous collaborative learning make collaborative learning inefficient in real-world heterogeneous environments. We propose a novel and efficient collaborative learning framework named AdaptCL, which generates an adaptive sub-model dynamically from the global base model for each data holder, without any prior information about worker capability. All workers (data holders) achieve approximately identical update time as the fastest worker by equipping them with capability-adapted pruned models. Thus the training process can be dramatically accelerated. Besides, we tailor the efficient pruned rate learning algorithm and pruning approach for AdaptCL. Meanwhile, AdaptCL provides a mechanism for handling the trade-off between accuracy and time overhead and can be combined with other techniques to accelerate training further. Empirical results show that AdaptCL introduces little computing and communication overhead. AdaptCL achieves time savings of more than 41\% on average and improves accuracy in a low heterogeneous environment. In a highly heterogeneous environment, AdaptCL achieves a training speedup of 6.2x with a slight loss of accuracy.
翻译:在多党合作学习中,参数服务器向每个数据持有人发送了一个全球模型,供当地培训使用,然后汇总全球承诺模式,以实现隐私保护;然而,同步协作学习的拖动问题和无同步协作学习的僵化问题使得合作学习在现实世界的多元环境中效率低下;我们提议了一个创新和有效的合作学习框架,名为 " 适应CL ",它从全球数据库模型中为每个数据持有人产生一个适应性的子模型,没有事先关于工人能力的任何信息。所有工人(数据持有者)通过为他们配备能力适应的修饰型模型,实现与最快的工人大致相同的更新时间。因此,培训过程可以大大加快。此外,我们为适应CLC调整高效的修剪速率学习算法和修剪方法。同时,DandCLL提供了处理准确度和时间管理之间的权衡机制,可以与其他技术相结合,进一步加速培训。 " 经验性结果 " 显示,适应CLE " 引入了少量计算和通信间接费用。 " 适应CL " 实现平均超过41 ⁇ 的时间节约,提高低混合环境的准确度。