The fine-grained relationship between form and function with respect to deep neural network architecture design and hardware-specific acceleration is one area that is not well studied in the research literature, with form often dictated by accuracy as opposed to hardware function. In this study, a comprehensive empirical exploration is conducted to investigate the impact of deep neural network architecture design on the degree of inference speedup that can be achieved via hardware-specific acceleration. More specifically, we empirically study the impact of a variety of commonly used macro-architecture design patterns across different architectural depths through the lens of OpenVINO microprocessor-specific and GPU-specific acceleration. Experimental results showed that while leveraging hardware-specific acceleration achieved an average inference speed-up of 380%, the degree of inference speed-up varied drastically depending on the macro-architecture design pattern, with the greatest speedup achieved on the depthwise bottleneck convolution design pattern at 550%. Furthermore, we conduct an in-depth exploration of the correlation between FLOPs requirement, level 3 cache efficacy, and network latency with increasing architectural depth and width. Finally, we analyze the inference time reductions using hardware-specific acceleration when compared to native deep learning frameworks across a wide variety of hand-crafted deep convolutional neural network architecture designs as well as ones found via neural architecture search strategies. We found that the DARTS-derived architecture to benefit from the greatest improvement from hardware-specific software acceleration (1200%) while the depthwise bottleneck convolution-based MobileNet-V2 to have the lowest overall inference time of around 2.4 ms.
翻译:与深神经网络结构设计和硬件特定加速有关的形式和功能之间的细微关系,是研究文献中未充分研究的一个领域,其形式往往取决于精确度,而不是硬件功能。在本研究中,进行了全面的实证探索,以调查深神经网络结构设计对通过硬件特定加速可以达到的推断速度程度的影响。更具体地说,我们通过OpenVINO微处理器特定和GPU特定加速度的透镜,对不同建筑深度的各种常用宏观结构设计模式的影响进行了实证研究。实验结果表明,在硬件特定加速度达到平均推导速度380 % 的同时,由于宏观结构设计模式的变化程度大不相同,在更深的瓶状卷内设计模式上实现的最大速度为550 % 。此外,我们深入探索了不同建筑深度和宽度结构的FLOP要求、3级缓存效率以及网络内嵌度和不断提高的建筑深度和宽度之间的关联性。最后,我们用最深层次的硬度结构来分析,在最深层次结构中,我们用最深层次的硬度结构来分析,在最深层次结构中,在最深层次结构中发现了时间结构中发现,在最深层次结构中发现了结构中发现了结构中找到了结构中发现,在不断降低。