Deep neural networks provide state-of-the-art accuracy for vision tasks but they require significant resources for training. Thus, they are trained on cloud servers far from the edge devices that acquire the data. This issue increases communication cost, runtime and privacy concerns. In this study, a novel hierarchical training method for deep neural networks is proposed that uses early exits in a divided architecture between edge and cloud workers to reduce the communication cost, training runtime and privacy concerns. The method proposes a brand-new use case for early exits to separate the backward pass of neural networks between the edge and the cloud during the training phase. We address the issues of most available methods that due to the sequential nature of the training phase, cannot train the levels of hierarchy simultaneously or they do it with the cost of compromising privacy. In contrast, our method can use both edge and cloud workers simultaneously, does not share the raw input data with the cloud and does not require communication during the backward pass. Several simulations and on-device experiments for different neural network architectures demonstrate the effectiveness of this method. It is shown that the proposed method reduces the training runtime by 29% and 61% in CIFAR-10 classification experiment for VGG-16 and ResNet-18 when the communication with the cloud is done at a low bit rate channel. This gain in the runtime is achieved whilst the accuracy drop is negligible. This method is advantageous for online learning of high-accuracy deep neural networks on low-resource devices such as mobile phones or robots as a part of an edge-cloud system, making them more flexible in facing new tasks and classes of data.
翻译:通过早期退出使用深度神经网络层次化训练
Deep neural networks provide state-of-the-art accuracy for vision tasks but they require significant resources for training. Thus, they are trained on cloud servers far from the edge devices that acquire the data. This issue increases communication cost, runtime and privacy concerns. In this study, a novel hierarchical training method for deep neural networks is proposed that uses early exits in a divided architecture between edge and cloud workers to reduce the communication cost, training runtime and privacy concerns. The method proposes a brand-new use case for early exits to separate the backward pass of neural networks between the edge and the cloud during the training phase. We address the issues of most available methods that due to the sequential nature of the training phase, cannot train the levels of hierarchy simultaneously or they do it with the cost of compromising privacy. In contrast, our method can use both edge and cloud workers simultaneously, does not share the raw input data with the cloud and does not require communication during the backward pass. Several simulations and on-device experiments for different neural network architectures demonstrate the effectiveness of this method. It is shown that the proposed method reduces the training runtime by 29% and 61% in CIFAR-10 classification experiment for VGG-16 and ResNet-18 when the communication with the cloud is done at a low bit rate channel. This gain in the runtime is achieved whilst the accuracy drop is negligible. This method is advantageous for online learning of high-accuracy deep neural networks on low-resource devices such as mobile phones or robots as a part of an edge-cloud system, making them more flexible in facing new tasks and classes of data.
深度神经网络为视觉任务提供了最先进的准确率,但它们需要大量的培训资源。因此,它们在远离获取数据的边缘设备的云服务器上进行训练。这个问题增加了通信成本、运行时和隐私问题。本研究提出了一种新的深度神经网络层次化训练方法,该方法在边缘和云工作者之间使用早期退出的划分结构,以减少通信成本、训练运行时间和隐私问题。该方法提出了一个全新的早期退出用例,用于在训练阶段将神经网络的反向传递分离在边缘和云之间。我们解决了大多数可用方法的问题,由于训练阶段的顺序性质,不能同时训练层次结构,或者为此付出安全权衡的代价。相反,我们的方法可以同时使用边缘和云工作者,不与云共享原始输入数据,不需要反向传递期间的通信。不同神经网络架构的几个模拟和设备实验证明了该方法的有效性。研究表明,当在低比特率信道上与云通信时,所提出的方法将CIFAR-10分类实验的VGG-16和ResNet-18的培训运行时间分别减少了29%和61%。这种运行时间的提高是在准确率下降可忽略的情况下实现的。该方法对于在低资源设备上进行高准确度深度神经网络的在线学习是有利的,例如移动电话或机器人作为边缘云系统的一部分,使它们更灵活地面对新任务和数据类别。