Deep neural network (DNN) inference is increasingly being executed on mobile and embedded platforms due to low latency and better privacy. However, efficient deployment on these platforms is challenging due to the intensive computation and memory access. We propose a holistic system design for DNN performance and energy optimisation, combining the trade-off opportunities in both algorithms and hardware. The system can be viewed as three abstract layers: the device layer contains heterogeneous computing resources; the application layer has multiple concurrent workloads; and the runtime resource management layer monitors the dynamically changing algorithms' performance targets as well as hardware resources and constraints, and tries to meet them by tuning the algorithm and hardware at the same time. Moreover, We illustrate the runtime approach through a dynamic version of 'once-for-all network' (namely Dynamic-OFA), which can scale the ConvNet architecture to fit heterogeneous computing resources efficiently and has good generalisation for different model architectures such as Transformer. Compared to the state-of-the-art Dynamic DNNs, our experimental results using ImageNet on a Jetson Xavier NX show that the Dynamic-OFA is up to 3.5x (CPU), 2.4x (GPU) faster for similar ImageNet Top-1 accuracy, or 3.8% (CPU), 5.1% (GPU) higher accuracy at similar latency. Furthermore, compared with Linux governor (e.g. performance, schedutil), our runtime approach reduces the energy consumption by 16.5% at similar latency.
翻译:深心神经网络(DNN) 的推论正在越来越多地在移动平台和嵌入平台上执行,原因是低延迟和更好的隐私。然而,由于大量计算和内存访问,在这些平台上的有效部署具有挑战性。我们提议了DNN性能和能源优化的综合系统设计,将算法和硬件的权衡机会结合起来。该系统可被视为三个抽象层:设备层包含多种计算资源;应用层有多重并行工作量;运行时间资源管理层监测动态变化算法的性能目标以及硬件资源和限制,并试图通过同时调整算法和硬件来达到它们。此外,我们通过动态版本的“全方位网络”性能和能源优化来说明运行时间方法。这个系统可以扩大ConvNet结构,以适应各种混合计算资源,并且能够很好地概括诸如变压器等不同的模型结构。与目前状态的动态动态 DNNNNNNNPU方法以及硬件和限制相比,我们使用图像网络的实验结果,同时调整算法和硬件的算法和硬件。此外,SDV-G-G-POVA的精确度(S 2.5) 和S-POVA的类似图像的精确度为35。