Adder neural networks (AdderNets) have shown impressive performance on image classification with only addition operations, which are more energy efficient than traditional convolutional neural networks built with multiplications. Compared with classification, there is a strong demand on reducing the energy consumption of modern object detectors via AdderNets for real-world applications such as autonomous driving and face detection. In this paper, we present an empirical study of AdderNets for object detection. We first reveal that the batch normalization statistics in the pre-trained adder backbone should not be frozen, since the relatively large feature variance of AdderNets. Moreover, we insert more shortcut connections in the neck part and design a new feature fusion architecture for avoiding the sparse features of adder layers. We present extensive ablation studies to explore several design choices of adder detectors. Comparisons with state-of-the-arts are conducted on COCO and PASCAL VOC benchmarks. Specifically, the proposed Adder FCOS achieves a 37.8\% AP on the COCO val set, demonstrating comparable performance to that of the convolutional counterpart with an about $1.4\times$ energy reduction.
翻译:添加神经网络(AdderNets)在图像分类上表现出了令人印象深刻的性能,只是增加了一些操作,这些操作比以倍增方式建立的传统进化神经网络更能提高能源效率。与分类相比,人们强烈要求减少通过AdderNets为现实世界应用而使用的现代物体探测器的能源消耗,例如自动驾驶和面部探测。我们在本文件中介绍了对AderNets的实验性研究,以探测物体。我们首先发现,由于AderNets具有相对较大的特点差异,培训前添加器骨干中的批量正常化统计数据不应被冻结。此外,我们在颈部插入了更多的捷径连接,并设计了一个新的特征聚合结构,以避免添加器层的稀少特征。我们提出了广泛的模拟研究,以探讨添加器探测器的若干设计选择。与COCO和PASAL VOC基准的状态比较。具体地说,拟议的Adder FCOS在COval集中实现了37.8 ⁇ AP,表明其性能与CCOVAL值相当于14美元能源削减的同量。