In this paper, we propose a simple and general approach to augment regular convolution operator by injecting extra group-wise transformation during training and recover it during inference. Extra transformation is carefully selected to ensure it can be merged with regular convolution in each group and will not change the topological structure of regular convolution during inference. Compared with regular convolution operator, our approach (AugConv) can introduce larger learning capacity to improve model performance during training but will not increase extra computational overhead for model deployment. Based on ResNet, we utilize AugConv to build convolutional neural networks named AugResNet. Result on image classification dataset Cifar-10 shows that AugResNet outperforms its baseline in terms of model performance.
翻译:在本文中,我们提出了一个简单和一般的办法,通过在培训期间注入额外的群体变换方法来增加正常的变换操作员,并在推论期间加以恢复。我们仔细选择了额外变换,以确保它能够与每个群体的正常变换合并,并且在推论期间不会改变正常变换的地形结构。与正常变换操作员相比,我们的方法(AugConv)可以引入更大的学习能力,以提高培训期间的模型性能,但不会增加模型部署的计算间接费用。根据ResNet,我们利用AugConv建立称为AugResNet的变换神经网络。关于图像分类数据集Cifar-10的结果显示,AugResNet在模型性能方面超过了基线。