Batch Normalization (BN) is an important preprocessing step to many deep learning applications. Since it is a data-dependent process, for some homogeneous datasets it is a redundant or even a performance-degrading process. In this paper, we propose an early-stage feasibility assessment method for estimating the benefits of applying BN on the given data batches. The proposed method uses a novel threshold-based approach to classify the training data batches into two sets according to their need for normalization. The need for normalization is decided based on the feature heterogeneity of the considered batch. The proposed approach is a pre-training processing, which implies no training overhead. The evaluation results show that the proposed approach achieves better performance mostly in small batch sizes than the traditional BN using MNIST, Fashion-MNIST, CIFAR-10, and CIFAR-100 datasets. Additionally, the network stability is increased by reducing the occurrence of internal variable transformation.
翻译:批量正常化(BN)是许多深层学习应用程序的一个重要预处理步骤,由于它是一个依赖数据的过程,因此对于一些同质数据集来说,它是一个多余的,甚至是一个性能退化的过程;在本文件中,我们提出一个早期的可行性评估方法,用以估计在给定数据批次上应用BN的好处;拟议方法采用基于新颖门槛的方法,将培训数据批次按其正常化需要分为两组;根据所考虑的批次的特征差异性决定正常化的必要性;拟议方法是一种培训前处理,不意味着培训间接费用;评价结果表明,拟议的方法比传统的BN少批量地使用MNIST、Fashon-MNIST、CIFAR-10和CIFAR-100数据集实现更好的业绩;此外,网络稳定性通过减少内部变异的发生而提高。