Learned classifiers should often possess certain invariance properties meant to encourage fairness, robustness, or out-of-distribution generalization. However, multiple recent works empirically demonstrate that common invariance-inducing regularizers are ineffective in the over-parameterized regime, in which classifiers perfectly fit (i.e. interpolate) the training data. This suggests that the phenomenon of ``benign overfitting," in which models generalize well despite interpolating, might not favorably extend to settings in which robustness or fairness are desirable. In this work we provide a theoretical justification for these observations. We prove that -- even in the simplest of settings -- any interpolating learning rule (with arbitrarily small margin) will not satisfy these invariance properties. We then propose and analyze an algorithm that -- in the same setting -- successfully learns a non-interpolating classifier that is provably invariant. We validate our theoretical observations on simulated data and the Waterbirds dataset.
翻译:学分分类者通常应拥有某些旨在鼓励公平、稳健或分配范围外一般化的差值属性。 但是,最近多项工作的经验证明,在过度参数化的制度中,常见的惯性诱导正规化者是无效的,因为分类者完全适合(即内插)培训数据。这表明“比重超编”现象,即模型尽管有内插作用,但普遍化的“比重超编”现象,可能不会有利地扩大到适宜稳健或公平性的环境。在这项工作中,我们为这些观察提供了理论依据。我们证明,即使在最简单的情况下,任何内插式学习规则(以任意小的幅度)都无法满足这些反常性特征。我们然后提出和分析一种算法,即在同一环境中,成功地学会了一种非内插式分类者,这种算法是容易变化的。我们验证了我们对模拟数据和水鸟数据集的理论意见。