A central goal of machine learning is to learn robust representations that capture the causal relationship between inputs features and output labels. However, minimizing empirical risk over finite or biased datasets often results in models latching on to spurious correlations between the training input/output pairs that are not fundamental to the problem at hand. In this paper, we define and analyze robust and spurious representations using the information-theoretic concept of minimal sufficient statistics. We prove that even when there is only bias of the input distribution (i.e. covariate shift), models can still pick up spurious features from their training data. Group distributionally robust optimization (DRO) provides an effective tool to alleviate covariate shift by minimizing the worst-case training loss over a set of pre-defined groups. Inspired by our analysis, we demonstrate that group DRO can fail when groups do not directly account for various spurious correlations that occur in the data. To address this, we further propose to minimize the worst-case losses over a more flexible set of distributions that are defined on the joint distribution of groups and instances, instead of treating each group as a whole at optimization time. Through extensive experiments on one image and two language tasks, we show that our model is significantly more robust than comparable baselines under various partitions. Our code is available at https://github.com/violet-zct/group-conditional-DRO.
翻译:机器学习的一个中心目标是学习能够捕捉投入特征和产出标签之间因果关系的稳健表述;然而,将有限或偏差数据集的经验风险降到最低,往往导致模型在对手头的问题来说并非根本的培训投入/产出对等之间出现虚假的关联。在本文中,我们利用信息理论概念,即最低限度的充分统计数据来定义和分析稳健和虚假的表述。我们证明,即使输入分布存在偏差(即可变性转换),模型仍然可以从其培训数据中获得虚假的特征。集体分布稳健优化(DRO)提供了有效的工具,通过在一组预先界定的组别中尽量减少最坏情况的培训损失来缓解共变。我们的分析启发了我们的分析,我们证明,如果各组没有直接说明数据中出现的各种虚假关联,DRO可以失败。为了解决这一问题,我们进一步建议尽量减少最坏的情况损失,因为一组和一组分布更加灵活,而不是将每个组类别视为一个组别,而不是将每个组类别视为一个大大稳健的基线。我们在两个基底线上展示了我们两个可比的模型。