With the continuous development of neural networks in computer vision tasks, more and more network architectures have achieved outstanding success. As one of the most advanced neural network architectures, DenseNet shortcuts all feature maps to solve the problem of model depth. Although this network architecture has excellent accuracy at low MACs (multiplications and accumulations), it takes excessive inference time. To solve this problem, HarDNet reduces the connections between feature maps, making the remaining connections resemble harmonic waves. However, this compression method may result in decreasing model accuracy and increasing MACs and model size. This network architecture only reduces the memory access time, its overall performance still needs to be improved. Therefore, we propose a new network architecture using threshold mechanism to further optimize the method of connections. Different numbers of connections for different convolutional layers are discarded to compress the feature maps in ThreshNet. The proposed network architecture used three datasets, CIFAR-10, CIFAR-100, and SVHN, to evaluate the performance for image classifications. Experimental results show that ThreshNet achieves up to 60% reduction in inference time compared to DenseNet, and up to 35% faster training speed and 20% reduction in error rate compared to HarDNet on these datasets.
翻译:随着计算机视觉任务中神经网络的持续发展,越来越多的网络结构取得了显著的成功。作为最先进的神经网络结构之一,DenseNet捷径将所有功能地图都用于解决模型深度问题。虽然这个网络结构在低 MACs(倍数和累积数)上非常精准,但需要过长的推算时间。为了解决这个问题,HarDNet减少了地貌图之间的连接,使其余连接与和谐波相类似。然而,这种压缩方法可能导致模型精度下降,摩卡和模型尺寸增加。这个网络结构只缩短了记忆存取时间,其总体性能仍有待改进。因此,我们提出一个新的网络结构,利用门槛机制进一步优化连接方法。不同革命层的不同连接数量被丢弃,以压缩ThreshNet的地貌图。拟议的网络结构使用了三个数据集,即CIFAR-10、CIFAR-100和SVHN,以评价图像分类的性能。实验结果表明,ScheshNet在与DenseNet数据率相比,降为60%,比HarseNet更快地降低了数据速度,降为35。