Large deep learning models have shown great potential for delivering exceptional results in various applications. However, the training process can be incredibly challenging due to the models' vast parameter sizes, often consisting of hundreds of billions of parameters. Common distributed training methods, such as data parallelism, tensor parallelism, and pipeline parallelism, demand significant data communication throughout the process, leading to prolonged wait times for some machines in physically distant distributed systems. To address this issue, we propose a novel solution called Hulk, which utilizes a modified graph neural network to optimize distributed computing systems. Hulk not only optimizes data communication efficiency between different countries or even different regions within the same city, but also provides optimal distributed deployment of models in parallel. For example, it can place certain layers on a machine in a specific region or pass specific parameters of a model to a machine in a particular location. By using Hulk in experiments, we were able to improve the time efficiency of training large deep learning models on distributed systems by more than 20\%. Our open source collection of unlabeled data:https://github.com/DLYuanGod/Hulk.
翻译:大型深度学习模型在各种应用中展现出了巨大的潜力,然而模型的训练过程由于其庞大的参数数量(通常包含数百亿个参数)而变得异常艰辛。常用的分布式训练方法,例如数据并行、张量并行和管道并行,在整个过程中需要进行大量的数据交流,导致分布式系统中某些机器的等待时间延长,尤其是在物理上分离的地区。为了解决这个问题,我们提出了一种名为绿巨人(Hulk)的全新解决方案。绿巨人利用改进过的图神经网络来优化分布式计算系统,不仅可以优化不同国家甚至同一城市内不同地区之间的数据传输效率,而且可以在并行计算中提供最佳分布式部署。例如,它可以将某些层放置在特定地区的计算机上,或将模型的特定参数传递给特定位置的机器。通过在实验中使用绿巨人,我们成功将大型深度学习模型在分布式系统上的训练时间效率提高了20%以上。我们的未标记数据开源收集:https://github.com/DLYuanGod/Hulk.