The rapid advancements of computing technology facilitate the development of diverse deep learning applications. Unfortunately, the efficiency of parallel computing infrastructures varies widely with neural network models, which hinders the exploration of the design space to find high-performance neural network architectures on specific computing platforms for a given application. To address such a challenge, we propose a deep learning-based method, ResPerfNet, which trains a residual neural network with representative datasets obtained on the target platform to predict the performance for a deep neural network. Our experimental results show that ResPerfNet can accurately predict the execution time of individual neural network layers and full network models on a variety of platforms. In particular, ResPerfNet achieves 8.4% of mean absolute percentage error for LeNet, AlexNet and VGG16 on the NVIDIA GTX 1080Ti, which is substantially lower than the previously published works.
翻译:不幸的是,平行计算机基础设施的效率与神经网络模型有很大差异,这阻碍了对设计空间的探索,以寻找特定应用特定计算机平台上高性能神经网络结构。为了应对这一挑战,我们建议采用深层次的基于学习的方法ResPerfNet(ResPerfNet),用目标平台上获得的代表性数据集对残余神经网络进行培训,以预测深神经网络的性能。我们的实验结果表明,ResPerfNet(ResPerfNet)可以准确预测各个神经网络层和各种平台上完整网络模型的执行时间。特别是,ResPerfNet(ResPerfNet)在LeNet(LeNet)、AlexNet(AlexNet)和VGG16(VGG16)方面实现了8.4%的绝对百分比误差,这一误差大大低于先前出版的著作。