Optimization of Top-1 ImageNet promotes enormous networks that may be impractical in inference settings. Binary neural networks (BNNs) have the potential to significantly lower the compute intensity but existing models suffer from low quality. To overcome this deficiency, we propose PokeConv, a binary convolution block which improves quality of BNNs by techniques such as adding multiple residual paths, and tuning the activation function. We apply it to ResNet-50 and optimize ResNet's initial convolutional layer which is hard to binarize. We name the resulting network family PokeBNN. These techniques are chosen to yield favorable improvements in both top-1 accuracy and the network's cost. In order to enable joint optimization of the cost together with accuracy, we define arithmetic computation effort (ACE), a hardware- and energy-inspired cost metric for quantized and binarized networks. We also identify a need to optimize an under-explored hyper-parameter controlling the binarization gradient approximation. We establish a new, strong state-of-the-art (SOTA) on top-1 accuracy together with commonly-used CPU64 cost, ACE cost and network size metrics. ReActNet-Adam, the previous SOTA in BNNs, achieved a 70.5% top-1 accuracy with 7.9 ACE. A small variant of PokeBNN achieves 70.5% top-1 with 2.6 ACE, more than 3x reduction in cost; a larger PokeBNN achieves 75.6% top-1 with 7.8 ACE, more than 5% improvement in accuracy without increasing the cost. PokeBNN implementation in JAX/Flax and reproduction instructions are available in AQT repository: https://github.com/google/aqt
翻译:优化 Top-1 图像网 的二进制会促进巨大的网络, 这些网络在推断设置中可能不切实际。 Binary 神经网络( BNNNS) 有可能大幅降低计算强度, 但现有模型的质量却低。 为了克服这一缺陷, 我们提议 Poke Conv, 这是一种二进制的共进式组合块, 通过添加多个剩余路径等技术来提高 BNNT 质量, 并调整激活功能。 我们将其应用到 ResNet- 50 并优化 ResNet 的初始循环层, 而 很难双进化。 我们命名了由此产生的网络家族 PokeBNNNNN。 选择了这些技术, 以在上一至一的精度和网络成本上更低。 为了能够联合优化成本和准确性, 我们定义了 poke- NEFCO 的计算方法, 一个硬件和能源激励成本衡量标准, 将A- NEO- NE- NEO 的高级成本比 AS A- NE- NEA 70。