As deep neural networks (DNNs) become deeper, the training time increases. In this perspective, multi-GPU parallel computing has become a key tool in accelerating the training of DNNs. In this paper, we introduce a novel methodology to construct a parallel neural network that can utilize multiple GPUs simultaneously from a given DNN. We observe that layers of DNN can be interpreted as the time step of a time-dependent problem and can be parallelized by emulating a parallel-in-time algorithm called parareal. The parareal algorithm consists of fine structures which can be implemented in parallel and a coarse structure which gives suitable approximations to the fine structures. By emulating it, the layers of DNN are torn to form a parallel structure which is connected using a suitable coarse network. We report accelerated and accuracy-preserved results of the proposed methodology applied to VGG-16 and ResNet-1001 on several datasets.
翻译:随着深神经网络(DNN)的深度加深,培训时间会增加。从这个角度看,多GPU平行计算已成为加速DNN培训的关键工具。在本文中,我们引入了一种新的方法来建造平行神经网络,可以同时利用特定DNN的多个GPU。我们观察到,DNN的层层可以被解释为一个时间依赖问题的时间步骤,并可以通过模拟一个称为假的平行时间算法来平行。 模拟算法包括一些精细结构,可以平行实施,而粗略结构可以使精细结构具有适当的近似值。通过模拟,DNNN将形成一个平行结构,利用一个合适的粗糙网络连接。我们报告几个数据集对VGG-16和ResNet-1001应用的拟议方法的加速和准确性结果。