Approaches based on deep neural networks have achieved striking performance when testing data and training data share similar distribution, but can significantly fail otherwise. Therefore, eliminating the impact of distribution shifts between training and testing data is crucial for building performance-promising deep models. Conventional methods assume either the known heterogeneity of training data (e.g. domain labels) or the approximately equal capacities of different domains. In this paper, we consider a more challenging case where neither of the above assumptions holds. We propose to address this problem by removing the dependencies between features via learning weights for training samples, which helps deep models get rid of spurious correlations and, in turn, concentrate more on the true connection between discriminative features and labels. Extensive experiments clearly demonstrate the effectiveness of our method on multiple distribution generalization benchmarks compared with state-of-the-art counterparts. Through extensive experiments on distribution generalization benchmarks including PACS, VLCS, MNIST-M, and NICO, we show the effectiveness of our method compared with state-of-the-art counterparts.
翻译:以深神经网络为基础的方法在测试数据和培训数据分布相似时取得了惊人的成绩,但也可能大失所望。因此,消除培训和测试数据之间的分布变化的影响对于建立符合业绩要求的深层模型至关重要。常规方法假定已知的培训数据(如域标签)的异质性或不同领域的大致同等能力。在本文件中,我们考虑上述假设均不存在的更具挑战性的情况。我们提议通过消除培训样本学习重量特征之间的依赖性来解决这一问题,这有助于深层模型摆脱虚假的关联,进而更多地注重歧视特征和标签之间的真正联系。广泛的实验清楚地表明了我们与最新对应方相比,多分布通用基准方法的有效性。通过对分布通用基准的广泛实验,包括PACS、VLCS、MNIST-M和NICO,我们展示了我们与最新对应方相比方法的有效性。