Convolutional neural networks (CNNs) are able to attain better visual recognition performance than fully connected neural networks despite having much less parameters due to their parameter sharing principle. Hence, modern architectures are designed to contain a very small number of fully-connected layers, often at the end, after multiple layers of convolutions. It is interesting to observe that we can replace large fully-connected layers with relatively small groups of tiny matrices applied on the entire image. Moreover, although this strategy already reduces the number of parameters, most of the convolutions can be eliminated as well, without suffering any loss in recognition performance. However, there is no solid recipe to detect this hidden subset of convolutional neurons that is responsible for the majority of the recognition work. Hence, in this work, we use the matrix characteristics based on eigenvalues in addition to the classical weight-based importance assignment approach for pruning to shed light on the internal mechanisms of a widely used family of CNNs, namely residual neural networks (ResNets), for the image classification problem using CIFAR-10, CIFAR-100 and Tiny ImageNet datasets.
翻译:与完全连接的神经网络相比,革命神经网络(CNNs)能够取得更好的视觉识别性能,尽管其参数共享原则的参数要少得多。因此,现代建筑的设计是包含极少数完全连接的层层,往往是在多层相融合之后的末端。有趣的是,我们可以用在整个图像上应用的相对小的微小基体来取代大连接层。此外,尽管这一战略已经减少了参数的数量,但大多数神经网络也可以消除,而不会在确认性能方面遭受任何损失。然而,没有可靠的方法来检测应对大部分识别工作负责的这一隐藏的卷子共生神经。因此,在这项工作中,我们除了使用传统的基于重量的重要分配方法外,还使用基于乙基值的矩阵特征特征,对广泛使用的CNN家族的内部机制,即残余神经网络(ResNets)进行搜索,以了解使用CIFAR-10、CIFAR-100和小图像网络数据集的图像分类问题。