We study the problem of out-of-distribution (o.o.d.) generalization where spurious correlations of attributes vary across training and test domains. This is known as the problem of correlation shift and has posed concerns on the reliability of machine learning. In this work, we introduce the concepts of direct and indirect effects from causal inference to the domain generalization problem. We argue that models that learn direct effects minimize the worst-case risk across correlation-shifted domains. To eliminate the indirect effects, our algorithm consists of two stages: in the first stage, we learn an indirect-effect representation by minimizing the prediction error of domain labels using the representation and the class labels; in the second stage, we remove the indirect effects learned in the first stage by matching each data with another data of similar indirect-effect representation but of different class labels in the training and validation phase. Our approach is shown to be compatible with existing methods and improve the generalization performance of them on correlation-shifted datasets. Experiments on 5 correlation-shifted datasets and the DomainBed benchmark verify the effectiveness of our approach.
翻译:我们研究分配外(o.o.d.)的概括问题,因为培训和测试领域的属性的虚假相关性各不相同。这被称为关联性转变问题,对机器学习的可靠性提出了关切。在这项工作中,我们引入了因果推论产生的直接和间接影响概念,以至领域概括化问题。我们争辩说,那些了解直接效应的模型将相关变换领域的最坏风险降到最低程度。为了消除间接影响,我们的算法包括两个阶段:在第一阶段,我们学习间接效应的表示法,通过使用代表制和类标签尽可能减少域标签的预测错误;在第二阶段,我们通过将每个数据与培训和验证阶段的另一种类似间接效应数据相匹配,而不同等级的标记,消除在第一阶段学到的间接效应。我们的方法与现有方法相容,改进在相关变数据集上的通用性表现。关于5个相关变数据集的实验和DomainB基准验证了我们方法的有效性。</s>