Deploying deep neural networks on low-resource edge devices is challenging due to their ever-increasing resource requirements. Recent investigations propose multiplication-free neural networks to reduce computation and memory consumption. Shift neural network is one of the most effective tools towards these reductions. However, existing low-bit shift networks are not as accurate as their full precision counterparts and cannot efficiently transfer to a wide range of tasks due to their inherent design flaws. We propose DenseShift network that exploits the following novel designs. First, we demonstrate that the zero-weight values in low-bit shift networks are neither useful to the model capacity nor simplify the model inference. Therefore, we propose to use a zero-free shifting mechanism to simplify inference while increasing the model capacity. Second, we design a new metric to measure the weight freezing issue in training low-bit shift networks, and propose a sign-scale decomposition to improve the training efficiency. Third, we propose the low-variance random initialization strategy to improve the model's performance in transfer learning scenarios. We run extensive experiments on various computer vision and speech tasks. The experimental results show that DenseShift network significantly outperforms existing low-bit multiplication-free networks and can achieve competitive performance to the full-precision counterpart. It also exhibits strong transfer learning performance with no drop in accuracy.
翻译:在低资源边缘装置上部署深层神经网络具有挑战性,因为其资源需求不断增加。最近的调查显示,低位转换网络的零重量值对模型能力没有用处,也没有简化模型推断。因此,我们提议使用零位转换机制来简化推断,同时增加模型能力。第二,我们设计了新的衡量标准,以衡量低位转换网络培训中的重量冻结问题,并提议采用一个标志化分解,以提高培训效率。第三,我们提议采用低变量随机初始化战略,以提高模型在转移学习情景中的性能。我们广泛试验各种计算机视觉和语音任务。实验结果显示,DenseShifft网络在提高模型转换能力的同时,还大大超越了现有低位转换功能的竞争力。