Many state-of-the-art ML results have been obtained by scaling up the number of parameters in existing models. However, parameters and activations for such large models often do not fit in the memory of a single accelerator device; this means that it is necessary to distribute training of large models over multiple accelerators. In this work, we propose PipeDream-2BW, a system that supports memory-efficient pipeline parallelism. PipeDream-2BW uses a novel pipelining and weight gradient coalescing strategy, combined with the double buffering of weights, to ensure high throughput, low memory footprint, and weight update semantics similar to data parallelism. In addition, PipeDream-2BW automatically partitions the model over the available hardware resources, while respecting hardware constraints such as memory capacities of accelerators and interconnect topologies. PipeDream-2BW can accelerate the training of large GPT and BERT language models by up to 20$\times$ with similar final model accuracy.
翻译:通过扩大现有模型的参数数量,取得了许多最先进的ML结果。然而,这些大型模型的参数和激活往往不适合单一加速器装置的记忆;这意味着有必要通过多个加速器对大型模型进行培训。在这项工作中,我们提议了支持内存高效管道平行的系统PipeDream-2BW。PipeDream-2BW使用一种新型的管线和权重梯联结战略,加上权重的双重缓冲,以确保高吞吐量、低记忆足迹和重量更新与数据平行的语义。此外,PipeDream-2BW将模型自动分隔在现有的硬件资源之上,同时尊重硬件限制,例如加速器的记忆能力和连接表层。PipeDream-2BW可以加快大型GPT语言模型和BERT语言模型的培训,最高可达20美元,最后模型精确度类似。