Out-of-distribution (OOD) generalisation aims to build a model that can well generalise its learnt knowledge from source domains to an unseen target domain. However, current image classification models often perform poorly in the OOD setting due to statistically spurious correlations learning from model training. From causality-based perspective, we formulate the data generation process in OOD image classification using a causal graph. On this graph, we show that prediction P(Y|X) of a label Y given an image X in statistical learning is formed by both causal effect P(Y|do(X)) and spurious effects caused by confounding features (e.g., background). Since the spurious features are domain-variant, the prediction P(Y|X) becomes unstable on unseen domains. In this paper, we propose to mitigate the spurious effect of confounders using front-door adjustment. In our method, the mediator variable is hypothesized as semantic features that are essential to determine a label for an image. Inspired by capability of style transfer in image generation, we interpret the combination of the mediator variable with different generated images in the front-door formula and propose novel algorithms to estimate it. Extensive experimental results on widely used benchmark datasets verify the effectiveness of our method.
翻译:分布外( OOD) 概略( OOD) 的目的是构建一个模型,能够将其从源域到不可见的目标域所学的知识广泛归纳为普通知识。 然而,由于从模型培训中学习的统计假相关联性,当前图像分类模型在 OOOD 设置中往往表现不佳。 从因果关系的角度,我们用因果图绘制OOOD 图像分类的数据生成过程。在这个图上,我们显示,在统计学习中给Y的图像X标签预测P(Y ⁇ X)是由因果效应P(Y ⁇ do(X))和由混杂特征(例如背景)造成的虚假影响形成的。由于虚假特征是域变量,所以在 OOOD 设置中,预测P(Y ⁇ X) 在未知的域中变得不稳定。 在本文中,我们提议用前门调整来减轻粘结器的刺激效应。 在我们的方法中, 调控器的偏小于确定图像标签的关键的语义性特征。 受图像生成能力的启发, 我们用图像生成的风格转换能力来解释调解员变量与前门模型模型模型模型模型分析结果。