Though remarkable progress has been achieved in various vision tasks, deep neural networks still suffer obvious performance degradation when tested in out-of-distribution scenarios. We argue that the feature statistics (mean and standard deviation), which carry the domain characteristics of the training data, can be properly manipulated to improve the generalization ability of deep learning models. Common methods often consider the feature statistics as deterministic values measured from the learned features and do not explicitly consider the uncertain statistics discrepancy caused by potential domain shifts during testing. In this paper, we improve the network generalization ability by modeling the uncertainty of domain shifts with synthesized feature statistics during training. Specifically, we hypothesize that the feature statistic, after considering the potential uncertainties, follows a multivariate Gaussian distribution. Hence, each feature statistic is no longer a deterministic value, but a probabilistic point with diverse distribution possibilities. With the uncertain feature statistics, the models can be trained to alleviate the domain perturbations and achieve better robustness against potential domain shifts. Our method can be readily integrated into networks without additional parameters. Extensive experiments demonstrate that our proposed method consistently improves the network generalization ability on multiple vision tasks, including image classification, semantic segmentation, and instance retrieval. The code can be available at https://github.com/lixiaotong97/DSU.
翻译:尽管在各种愿景任务中取得了显著进展,但深神经网络在超出分配的情景中进行测试后仍然明显地出现性能退化。我们认为,带有培训数据领域特点的特征统计(平均值和标准偏差)可以适当地加以操纵,以提高深层学习模式的普遍化能力。常见方法往往将特征统计视为从所学特征中测量的决定性价值,而没有明确地考虑到测试期间可能发生的领域变化造成的不确定统计差异。在本文中,我们通过在培训期间以综合特征统计数据模拟域变的不确定性,提高网络的概括化能力。具体地说,我们假设,在考虑潜在不确定性之后,特征统计遵循多变量高斯分布。因此,每种特征统计不再是一个确定性价值,而是具有不同分布可能性的概率点。由于特征统计数据不确定,可以培训模型,以缓解域变变变变化的可能性,并针对潜在的域变更稳性。我们的方法可以在没有额外参数的情况下很容易融入网络。广泛的实验表明,我们提出的方法可以持续地改进网络的通用能力,在多种愿景任务上,包括Segi/Seximial am recal。