Deep classifiers are known to rely on spurious features $\unicode{x2013}$ patterns which are correlated with the target on the training data but not inherently relevant to the learning problem, such as the image backgrounds when classifying the foregrounds. In this paper we evaluate the amount of information about the core (non-spurious) features that can be decoded from the representations learned by standard empirical risk minimization (ERM) and specialized group robustness training. Following recent work on Deep Feature Reweighting (DFR), we evaluate the feature representations by re-training the last layer of the model on a held-out set where the spurious correlation is broken. On multiple vision and NLP problems, we show that the features learned by simple ERM are highly competitive with the features learned by specialized group robustness methods targeted at reducing the effect of spurious correlations. Moreover, we show that the quality of learned feature representations is greatly affected by the design decisions beyond the training method, such as the model architecture and pre-training strategy. On the other hand, we find that strong regularization is not necessary for learning high quality feature representations. Finally, using insights from our analysis, we significantly improve upon the best results reported in the literature on the popular Waterbirds, CelebA hair color prediction and WILDS-FMOW problems, achieving 97%, 92% and 50% worst-group accuracies, respectively.
翻译:据了解,深层分类者所依赖的是与培训数据目标相关但与学习问题本身无关的虚假特征 $\ unicode{x2013} 美元 模式,这些特征与培训数据目标相关,但与学习问题无关,例如对前台进行分类时的图像背景。在本文件中,我们评估了从标准风险最小化(ERM)和专门团体强力培训所吸取的表述中可以解码的核心(非净)特征的信息量。在近期关于深功能重配(DFR)的工作之后,我们通过再培训模型最后一层的特征表现来评估这些特征表现,在其中出现虚假相关性的封闭式数据集中,我们发现,在多个愿景和NLP问题上,我们发现简单的机构化所学到的特征与专门团体强力方法所学到的特征高度竞争,这些特征旨在减少虚假关联效应的影响。此外,我们表明,所学过的特征表现的质量受到培训方法以外的设计决定(如最差的架构和训练前战略)的极大影响。另一方面,我们发现,在学习高质量特征展示高质量特征展示时,没有必要进行优化的正规化。关于50级的图像分析,我们分别报告了SLISFMFDA的深度分析。