The existence of spurious correlations such as image backgrounds in the training environment can make empirical risk minimization (ERM) perform badly in the test environment. To address this problem, Kirichenko et al. (2022) empirically found that the core features that are causally related to the outcome can still be learned well even with the presence of spurious correlations. This opens a promising strategy to first train a feature learner rather than a classifier, and then perform linear probing (last layer retraining) in the test environment. However, a theoretical understanding of when and why this approach works is lacking. In this paper, we find that core features are only learned well when they are less noisy than spurious features, which is not necessarily true in practice. We provide both theories and experiments to support this finding and to illustrate the importance of feature noise. Moreover, we propose an algorithm called Freeze then Train (FTT), that first freezes certain salient features and then trains the rest of the features using ERM. We theoretically show that FTT preserves features that are more beneficial to test time probing. Across two commonly used real-world benchmarks, FTT outperforms ERM, JTT and CVaR-DRO, with especially substantial improvement in accuracy (by 4.8%) when the feature noise is large.
翻译:培训环境中的图像背景等虚假关联的存在,使得实验风险最小化(ERM)在测试环境中效果不佳。为了解决这一问题,Kirichenko等人(2022年)从经验上认为,即使存在虚假关联,与结果有因果关系的核心特征仍然可以很好地学习。这开启了一种有希望的战略,首先培训一个特征学习者而不是分类者,然后在测试环境中进行线性探测(最后一层再培训),然而,理论上理解这一方法何时和为什么没有奏效。在本文中,我们发现核心特征只有在不那么吵闹而不是虚假特征时才很好地学习。我们提供理论和实验,以支持这一发现,并表明特征噪音的重要性。此外,我们提议一种叫作“冻结”的算法,首先冻结某些突出特征,然后用机构风险管理对其余特征进行培训。我们理论上表明,FTT为测试时间保有更有利的特征。在通常使用的两种现实世界基准中,FTTT超出标准值,特别是标准值为4.8R的大幅改进时,JTTT和CVa的C值为4.8%。