Binary neural networks (BNNs), where both weights and activations are binarized into 1 bit, have been widely studied in recent years due to its great benefit of highly accelerated computation and substantially reduced memory footprint that appeal to the development of resource constrained devices. In contrast to previous methods tending to reduce the quantization error for training BNN structures, we argue that the binarized convolution process owns an increasing linearity towards the target of minimizing such error, which in turn hampers BNN's discriminative ability. In this paper, we re-investigate and tune proper non-linear modules to fix that contradiction, leading to a strong baseline which achieves state-of-the-art performance on the large-scale ImageNet dataset in terms of accuracy and training efficiency. To go further, we find that the proposed BNN model still has much potential to be compressed by making a better use of the efficient binary operations, without losing accuracy. In addition, the limited capacity of the BNN model can also be increased with the help of group execution. Based on these insights, we are able to improve the baseline with an additional 4~5% top-1 accuracy gain even with less computational cost. Our code will be made public at https://github.com/zhuogege1943/birealnetv2.
翻译:由于高速计算和大量减少记忆足迹对于开发资源受限装置具有巨大好处,近年来对加权和活化二进制神经网络(BNN)进行了广泛研究。与以往倾向于减少培训BNN结构的量化错误的方法相比,我们认为,二进制神经网络进程在尽量减少这种错误的目标方面具有越来越大的分界线性,这反过来又妨碍了BNN的偏向能力。在本文中,我们重新对适当的非线性模块进行重新投资并调制,以纠正这一矛盾,从而形成一个强大的基线,在大规模图像网络数据集中实现最先进的性能,精确度和培训效率。我们进一步发现,拟议的BNNN模型仍然有很大潜力通过更好地使用高效的二进制操作进行压缩,同时又不失去准确性。此外,BNN模型的有限能力也可以随着集体执行的帮助而增加。基于这些认识,我们有能力改进大规模图像网络数据网络数据库的基线,在准确度和培训效率方面达到最先进的性能度3/3。我们要发现,在4~5%/MARC 最高精确度上改进我们的公共代码。