Normalization techniques are crucial in stabilizing and accelerating the training of deep neural networks. However, they are mainly designed for the independent and identically distributed (IID) data, not satisfying many real-world out-of-distribution (OOD) situations. Unlike most previous works, this paper presents two normalization methods, SelfNorm and CrossNorm, to promote OOD generalization. SelfNorm uses attention to recalibrate statistics (channel-wise mean and variance), while CrossNorm exchanges the statistics between feature maps. SelfNorm and CrossNorm can complement each other in OOD generalization, though exploring different directions in statistics usage. Extensive experiments on different domains (vision and language), tasks (classification and segmentation), and settings (supervised and semi-supervised) show their effectiveness.
翻译:常规化技术对于稳定和加速深层神经网络的培训至关重要,但是,这些技术主要是为独立和相同分布的(IID)数据设计的,不能满足许多现实世界外分配(OOOD)情况,与大多数以前的工作不同,本文件介绍了两种常规化方法,即SelfNorm和CrossNorm,以促进OOOD一般化。自我诺姆利用对重新校正统计的注意(循环平均和差异),而CrossNorm在地貌地图之间交换统计数据。自我诺姆和CrossNorm在OOD一般化方面可以相互补充,尽管探索统计使用的不同方向。在不同领域(视觉和语言)、任务(分类和分类)和环境(监督和半监督)的广泛实验显示了其有效性。