Modern deep networks generally implement a certain form of shortcut connections to alleviate optimization difficulties. However, we observe that such network topology alters the nature of deep networks. In many ways these networks behave similarly to aggregated wide networks. We thus exploit the aggregation nature of shortcut connections at a finer architectural level and place them within wide convolutional layers. We end up with a sequentially aggregated convolutional (SeqConv) layer that combines the benefits of both wide and deep representations by aggregating features of various depths in sequence. The proposed SeqConv serves as a drop-in replacement of regular wide convolutional layers and thus could be handily integrated into any backbone network. We apply SeqConv to widely adopted backbones including ResNet and ResNeXt, and conduct experiments for image classification on public benchmark datasets. Our ResNet based network with a model size of ResNet-50 easily surpasses the performance of the 2.35$\times$ larger ResNet-152, while our ResNeXt based model sets a new state-of-the-art accuracy on ImageNet classification for networks with similar model complexity. The code and pretrained models of our work will be publicly released after review.
翻译:现代深层网络一般采用某种形式的捷径连接,以缓解优化困难。然而,我们观察到,此类网络地形学改变了深层网络的性质。在许多方面,这些网络的行为方式与广域网类似。我们因此利用了精细建筑层面的捷径连接总合性质,并将其置于大革命层中。我们最终出现了一个按顺序相加的连动(Seq Conv)层(Seq Convention)层,通过将各种深度的特征相加,将广泛和深层表达的好处结合在一起。拟议的Seq Conv 模式可以取代常规的广度电动层,从而可以方便地融入任何主干网络。我们将Seq Conv 应用于广泛采用的骨干,包括ResNet和ResNeXt,并在公共基准数据集上进行图像分类实验。我们基于ResNet-50模型的ResNet-50网络网络很容易超过2.35美元(times ResNet-152)的功能,而我们基于ResNeXt的模型则为具有类似复杂型号网络的图像网络的新的最新准确性。我们的工作代码和预设模式将在审查后公开发布。