Binary neural networks (BNNs) have received ever-increasing popularity for their great capability of reducing storage burden as well as quickening inference time. However, there is a severe performance drop compared with {real-valued} networks, due to its intrinsic frequent weight oscillation during training. In this paper, we introduce a Resilient Binary Neural Network (ReBNN) to mitigate the frequent oscillation for better BNNs' training. We identify that the weight oscillation mainly stems from the non-parametric scaling factor. To address this issue, we propose to parameterize the scaling factor and introduce a weighted reconstruction loss to build an adaptive training objective. %To the best of our knowledge, it is the first work to solve BNNs based on a dynamically re-weighted loss function. For the first time, we show that the weight oscillation is controlled by the balanced parameter attached to the reconstruction loss, which provides a theoretical foundation to parameterize it in back propagation. Based on this, we learn our ReBNN by {calculating} the {balanced} parameter {based on} its maximum magnitude, which can effectively mitigate the weight oscillation with a resilient training process. Extensive experiments are conducted upon various network models, such as ResNet and Faster-RCNN for computer vision, as well as BERT for natural language processing. The results demonstrate the overwhelming performance of our ReBNN over prior arts. For example, our ReBNN achieves 66.9\% Top-1 accuracy with ResNet-18 backbone on the ImageNet dataset, surpassing existing state-of-the-arts by a significant margin. Our code is open-sourced at https://github.com/SteveTsui/ReBNN.
翻译:在本文中,我们引入了弹性二进制神经网络(REBNN),以减少频繁的振动,以更好地进行BNN的训练。我们发现,重量振动主要来自非参数缩放系数。为解决这一问题,我们提议对缩放系数进行参数化,并引入加权重建损失,以建立适应性培训目标。根据我们的知识,这是在动态重新加权损失功能基础上解决BNN的首项工作。我们第一次展示了重力二进制双进制神经网络网络(REBNN),以缓解频繁的振动,以更好地进行更好的BNNNNNNN培训。我们发现,重振动主要来自非参数的缩放。基于此,我们通过调整缩放系数,引入一个加权的重力重建损失来构建一个适应性培训目标。我们网络的精度调整值比值(NNNFNB),这是我们最强的S-NEFRS-RRR) 快速的模型,可以有效地展示我们现有的图像模型。