Neural network binarization accelerates deep models by quantizing their weights and activations into 1-bit. However, there is still a huge performance gap between Binary Neural Networks (BNNs) and their full-precision (FP) counterparts. As the quantization error caused by weights binarization has been reduced in earlier works, the activations binarization becomes the major obstacle for further improvement of the accuracy. BNN characterises a unique and interesting structure, where the binary and latent FP activations exist in the same forward pass (\textit{i.e.} $\text{Binarize}(\mathbf{a}_F) = \mathbf{a}_B$). To mitigate the information degradation caused by the binarization operation from FP to binary activations, we establish a novel contrastive learning framework while training BNNs through the lens of Mutual Information (MI) maximization. MI is introduced as the metric to measure the information shared between binary and FP activations, which assists binarization with contrastive learning. Specifically, the representation ability of the BNNs is greatly strengthened via pulling the positive pairs with binary and FP activations from the same input samples, as well as pushing negative pairs from different samples (the number of negative pairs can be exponentially large). This benefits the downstream tasks, not only classification but also segmentation and depth estimation,~\textit{etc}. The experimental results show that our method can be implemented as a pile-up module on existing state-of-the-art binarization methods and can remarkably improve the performance over them on CIFAR-10/100 and ImageNet, in addition to the great generalization ability on NYUD-v2.
翻译:神经网络的二进制通过对重量进行量化并将其激活到 1 比位来加速深度模型。 但是, 二进制神经网络( BNNS) 和全精度( FP) 对等方之间仍然存在着巨大的性能差距。 由于早些时候的工程减少了由重量二进制( binar) 造成的量化错误, 启动二进制成为进一步提高准确性的主要障碍。 BNNS 描述一个独特和有趣的结构, 即二进制和潜化FPUD的激活存在于同一个远端传输器( textit{ i.e.} $\ text{Binalizeization} (mathbf{a{a{F) =\ gmathbf{ a ⁇ B$$。 由于从 FP的二进制操作到二进制( b进制), 我们建立了一个新的对比学习框架, 通过共同信息( MI) 最大化。 MI 是用来测量二进制和FP 之间共享信息的方法,, 也可以被引入, 双进制( 双进制) 通过对比学习, 硬制的机机能, 演示的演示的演示能, 。