In this paper, we propose several ideas for enhancing a binary network to close its accuracy gap from real-valued networks without incurring any additional computational cost. We first construct a baseline network by modifying and binarizing a compact real-valued network with parameter-free shortcuts, bypassing all the intermediate convolutional layers including the downsampling layers. This baseline network strikes a good trade-off between accuracy and efficiency, achieving superior performance than most of existing binary networks at approximately half of the computational cost. Through extensive experiments and analysis, we observed that the performance of binary networks is sensitive to activation distribution variations. Based on this important observation, we propose to generalize the traditional Sign and PReLU functions, denoted as RSign and RPReLU for the respective generalized functions, to enable explicit learning of the distribution reshape and shift at near-zero extra cost. Lastly, we adopt a distributional loss to further enforce the binary network to learn similar output distributions as those of a real-valued network. We show that after incorporating all these ideas, the proposed ReActNet outperforms all the state-of-the-arts by a large margin. Specifically, it outperforms Real-to-Binary Net and MeliusNet29 by 4.0% and 3.6% respectively for the top-1 accuracy and also reduces the gap to its real-valued counterpart to within 3.0% top-1 accuracy on ImageNet dataset. Code and models are available at: https://github.com/liuzechun/ReActNet.
翻译:在本文中,我们提出若干加强二进制网络的想法,以便在不产生任何额外的计算费用的情况下,消除实际价值网络的准确性差距。我们首先通过修改和采用无参数捷径,绕过所有中间革命层(包括下层抽样层),绕过所有中间革命层,绕过所有中间革命层(包括下层抽样层),构建一个基线网络。这一基线网络在准确性与效率之间取得了良好的平衡,以大约一半计算成本实现优于现有多数现有二进制网络的绩效。通过广泛的实验和分析,我们发现二进制网络的性能对激活分销变异十分敏感。基于这一重要观察,我们提议将传统的Sign和PRELU功能普遍化为无参数的精度,并将此称为Serign和RPRPRELU, 从而能够明确学习分布重塑和转换近零额外成本。最后,我们采用分配损失的方法进一步实施二进制网络,以学习与真实价值网络类似的产出分布。我们发现,在纳入所有这些想法后,拟议的ReAPNet将所有真实值/网络的准确性网络的准确性,将所有正态/正成正标值/正值数据在4.和Mel-irmaxxxxxxxxxxxxxxxxxxxxxxx