Normalization is an important but understudied challenge in privacy-related application domains such as federated learning (FL) and differential privacy (DP). While the unsuitability of batch normalization for FL and DP has already been shown, the impact of the other normalization methods on the performance of federated or differentially private models is not well-known. To address this, we draw a performance comparison among layer normalization (LayerNorm), group normalization (GroupNorm), and the recently proposed kernel normalization (KernelNorm) in FL and DP settings. Our results indicate LayerNorm and GroupNorm provide no performance gain compared to the baseline (i.e. no normalization) for shallow models, but they considerably enhance performance of deeper models. KernelNorm, on the other hand, significantly outperforms its competitors in terms of accuracy and convergence rate (or communication efficiency) for both shallow and deeper models. Given these key observations, we propose a kernel normalized ResNet architecture called KNResNet-13 for differentially private learning environments. Using the proposed architecture, we provide new state-of-the-art accuracy values on the CIFAR-10 and Imagenette datasets.
翻译:与隐私有关的应用领域,如联合学习(FL)和差异隐私(DP),正常化是一项重要但研究不足的挑战。虽然已经显示FL和DP分批正常化不适宜,但其他正常化方法对Federal化或差别化私人模型性能的影响并不广为人知。为了解决这个问题,我们比较了FL和DP环境中的层正常化(LayerNorm)、群体正常化(GroupNorm)和最近提议的内核正常化(KernelNorm)之间的性能比较。我们的结果显示,TelmNorm和GroupNorm没有提供与浅模型基线(即不正常化)相比的性能收益,但它们大大加强了更深模型的性能。另一方面,Kernelorm在浅度和更深层模型的准确率(或通信效率)方面大大超越了其竞争对手。根据这些关键观察,我们提议为差异私人学习环境而建立名为KNRRESNet-13的内核网络结构。我们利用拟议的结构,提供了新的州-FAR-10号图像-net精度值。