Batch Normalization is an essential component of all state-of-the-art neural networks architectures. However, since it introduces many practical issues, much recent research has been devoted to designing normalization-free architectures. In this paper, we show that weights initialization is key to train ResNet-like normalization-free networks. In particular, we propose a slight modification to the summation operation of a block output to the skip connection branch, so that the whole network is correctly initialized. We show that this modified architecture achieves competitive results on CIFAR-10 without further regularization nor algorithmic modifications.
翻译:批量正常化是所有最先进的神经网络结构的基本组成部分。 但是,由于它提出了许多实际问题,最近的许多研究都致力于设计无正常化结构。 在本文中,我们表明权重初始化是培训ResNet类似无正常化网络的关键。特别是,我们建议对区块输出与跳过连接分支的组合操作略作修改,以使整个网络得到正确的初始化。我们表明,这个经过修改的架构在没有进一步的正规化或算法修改的情况下,在CIFAR-10上取得了竞争性成果。