Deep neural networks (DNN) have achieved remarkable success in computer vision (CV). However, training and inference of DNN models are both memory and computation intensive, incurring significant overhead in terms of energy consumption and silicon area. In particular, inference is much more cost-sensitive than training because training can be done offline with powerful platforms, while inference may have to be done on battery powered devices with constrained form factors, especially for mobile or edge vision applications. In order to accelerate DNN inference, model quantization was proposed. However previous works only focus on the quantization rate without considering the efficiency of operations. In this paper, we propose Dendrite-Tree based Neural Network (DTNN) for energy-efficient inference with table lookup operations enabled by activation quantization. In DTNN both costly weight access and arithmetic computations are eliminated for inference. We conducted experiments on various kinds of DNN models such as LeNet-5, MobileNet, VGG, and ResNet with different datasets, including MNIST, Cifar10/Cifar100, SVHN, and ImageNet. DTNN achieved significant energy saving (19.4X and 64.9X improvement on ResNet-18 and VGG-11 with ImageNet, respectively) with negligible loss of accuracy. To further validate the effectiveness of DTNN and compare with state-of-the-art low energy implementation for edge vision, we design and implement DTNN based MLP image classifiers using off-the-shelf FPGAs. The results show that DTNN on the FPGA, with higher accuracy, could achieve orders of magnitude better energy consumption and latency compared with the state-of-the-art low energy approaches reported that use ASIC chips.
翻译:深心神经网络(DNN)在计算机愿景(CV)中取得了显著的成功。然而,DNN模型的培训和推论在存储和计算两方面都是密集的,在能源消耗和硅区域方面造成了巨大的间接成本。 特别是,由于培训可以使用强大的平台进行,因此成本比培训更加敏感,因为培训可以在离线上进行,而对于电池动力装置,特别是移动或边缘视觉应用,可能不得不作出推论。为了加快 DNNN的误判,提出了模型量化的模型。然而,先前的工作仅侧重于测量率,而没有考虑操作效率。在本文件中,我们提议以Dendret-Tree网络为基础的神经网络(DTNNNNNN)进行高能效的推断,因为启动石化平台,在DNFNNNW中, 高重量接入和算计算方法被排除。 我们用各种DNNNM模型进行了实验, 低端网络, 低边缘网络, VGGG和ResNet, 不同数据集,包括MNI、C10-CFDM-DDMS-DMS-DMS-DS-DDDDD 的升级, 和S-NGS-DDDS-DS-DS-NGS-DS-DS-S-DS-DS-DS-DDDS-DD SA的升级, SA-S-S-S-S-S-S-S-S-S-S-DDDDDDDDDD 和SDDSDDDDDDDDDDDDDDD 和S-S-S-D 和S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-SDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDD SA-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-