Current deep neural networks (DNNs) are overparameterized and use most of their neuronal connections during inference for each task. The human brain, however, developed specialized regions for different tasks and performs inference with a small fraction of its neuronal connections. We propose an iterative pruning strategy introducing a simple importance-score metric that deactivates unimportant connections, tackling overparameterization in DNNs and modulating the firing patterns. The aim is to find the smallest number of connections that is still capable of solving a given task with comparable accuracy, i.e. a simpler subnetwork. We achieve comparable performance for LeNet architectures on MNIST, and significantly higher parameter compression than state-of-the-art algorithms for VGG and ResNet architectures on CIFAR-10/100 and Tiny-ImageNet. Our approach also performs well for the two different optimizers considered -- Adam and SGD. The algorithm is not designed to minimize FLOPs when considering current hardware and software implementations, although it performs reasonably when compared to the state of the art.
翻译:目前的深神经网络(DNNs)被过度分解,在对每项任务进行推论时使用其大部分神经神经连接。然而,人类大脑为不同任务开发了专门区域,并对其神经连接的一小部分进行推论。我们提议了一项迭代裁剪战略,引入一个简单的、重要和核心的测量标准,将无关紧要的连接解除,在DNS中处理超分数,并调整发射模式。目的是找到仍然能够以类似精确度解决某项特定任务的最小数目的连接,即一个更简单的子网络。我们在MNIST上实现了LeNet结构的类似性能,并在CIFAR-10100和Tiny-ImaageNet上比最新的VGG和ResNet结构的参数压缩率要高得多。我们的方法对所审议的两个不同的优化者 -- -- Adam和SGD也表现良好。在考虑当前的硬件和软件实施时,算算法不是为了尽量减少FLOP,尽管与艺术状况相比,它的表现是合理的。