As designing appropriate Convolutional Neural Network (CNN) architecture in the context of a given application usually involves heavy human works or numerous GPU hours, the research community is soliciting the architecture-neutral CNN structures, which can be easily plugged into multiple mature architectures to improve the performance on our real-world applications. We propose Asymmetric Convolution Block (ACB), an architecture-neutral structure as a CNN building block, which uses 1D asymmetric convolutions to strengthen the square convolution kernels. For an off-the-shelf architecture, we replace the standard square-kernel convolutional layers with ACBs to construct an Asymmetric Convolutional Network (ACNet), which can be trained to reach a higher level of accuracy. After training, we equivalently convert the ACNet into the same original architecture, thus requiring no extra computations anymore. We have observed that ACNet can improve the performance of various models on CIFAR and ImageNet by a clear margin. Through further experiments, we attribute the effectiveness of ACB to its capability of enhancing the model's robustness to rotational distortions and strengthening the central skeleton parts of square convolution kernels.
翻译:由于在特定应用背景下设计适当的革命神经网络架构通常涉及大量的人类工程或无数的GPU小时,研究界正在征求建筑中立型CNN结构,这些结构可以很容易地插入多种成熟的建筑,以提高我们真实世界应用程序的性能。我们提议将非对称革命区(ACB)这一建筑中性结构作为CNN的建筑构件,它使用1D非对称性演动来强化平面演动核心网,对于现成的建筑,我们用ACB来取代标准平方圆层,以建设可接受更高精确度培训的AACNet(ACNet),在培训后,我们将ACNet等同地转换为相同的原始结构,因此不再需要额外计算。我们观察到ACNet可以明显地改善CRFAR和图像网络上各种模型的性能。通过进一步试验,我们把ACB的功效归功于其增强模型的稳健性,以达到轮换性,并加强CVAR的中央骨架骨骼部分。