Deep Convolutional Neural Networks (CNNs) have been widely used in various domains due to their impressive capabilities. These models are typically composed of a large number of 2D convolutional (Conv2D) layers with numerous trainable parameters. To reduce the complexity of a network, compression techniques can be applied. These methods typically rely on the analysis of trained deep learning models. However, in some applications, due to reasons such as particular data or system specifications and licensing restrictions, a pre-trained network may not be available. This would require the user to train a CNN from scratch. In this paper, we aim to find an alternative parameterization to Conv2D filters without relying on a pre-trained convolutional network. During the analysis, we observe that the effective rank of the vectorized Conv2D filters decreases with respect to the increasing depth in the network, which then leads to the implementation of the Depthwise Convolutional Eigen-Filter (DeCEF) layer. Essentially, a DeCEF layer is a low rank version of the Conv2D layer with significantly fewer trainable parameters and floating point operations (FLOPs). The way we define the effective rank is different from the previous work and it is easy to implement in any deep learning frameworks. To evaluate the effectiveness of DeCEF, experiments are conducted on the benchmark datasets CIFAR-10 and ImageNet using various network architectures. The results have shown a similar or higher accuracy and robustness using about 2/3 of the original parameters and reducing the number of FLOPs to 2/3 of the base network, which is then compared to the state-of-the-art techniques.
翻译:深层神经网络(CNNs)由于能力令人印象深刻,在不同领域被广泛使用。这些模型通常由大量2D的2D的进化(Conv2D)过滤器组成,具有许多可训练参数。为了降低网络的复杂性,可以应用压缩技术。这些方法通常依赖于对经过训练的深层学习模型的分析。然而,在某些应用中,由于特定的数据或系统规格和许可证限制等原因,可能没有经过预先培训的网络。这需要用户从头开始训练CNN。在本文中,我们的目标是为Conv2D过滤器找到替代参数的替代参数,而不必依赖事先训练过的进化网络。在分析中,我们观察到,随着网络深度的提高,Conv2D过滤器的有效级别,我们使用FLOP的原始级别和浮动基准框架来进行不同的评估。我们从原来的CFRFR的原始级别和浮动基准框架来定义实际的级别,在使用前FRFFF的模型和最易移动的模型上,我们使用了不同的数据框架。