Depth factorization and quantization have emerged as two of the principal strategies for designing efficient deep convolutional neural network (CNN) architectures tailored for low-power inference on the edge. However, there is still little detailed understanding of how different depth factorization choices affect the final, trained distributions of each layer in a CNN, particularly in the situation of quantized weights and activations. In this study, we introduce a progressive depth factorization strategy for efficient CNN architecture exploration under quantization constraints. By algorithmically increasing the granularity of depth factorization in a progressive manner, the proposed strategy enables a fine-grained, low-level analysis of layer-wise distributions. Thus enabling the gain of in-depth, layer-level insights on efficiency-accuracy tradeoffs under fixed-precision quantization. Such a progressive depth factorization strategy also enables efficient identification of the optimal depth-factorized macroarchitecture design (which we will refer to here as FactorizeNet) based on the desired efficiency-accuracy requirements.
翻译:深度因子化和量化是设计高效深电动神经网络结构的主要战略之一,这些结构是针对边缘低功率推断而专门设计的,然而,对于不同深度因子化选择如何影响有线电视新闻网中每一层最终、经过培训的分布,特别是在量化权重和激活的情况下,对于有线电视新闻网架构在量化限制下进行高效探索,我们提出了渐进深度因子化战略。通过循序渐进地增加深度因子化的颗粒,拟议战略能够对分层分布进行精细的、低层次的分析,从而能够在固定精确权分中获取关于效率-准确性权衡的深入、层次的洞察力。这种渐进深度因子化战略还有助于根据预期的效率-准确性要求,有效地确定最佳深度计分解的宏观结构设计(我们这里称为“因子化网 ” )。