Neural networks trained with SGD were recently shown to rely preferentially on linearly-predictive features and can ignore complex, equally-predictive ones. This simplicity bias can explain their lack of robustness out of distribution (OOD). The more complex the task to learn, the more likely it is that statistical artifacts (i.e. selection biases, spurious correlations) are simpler than the mechanisms to learn. We demonstrate that the simplicity bias can be mitigated and OOD generalization improved. We train a set of similar models to fit the data in different ways using a penalty on the alignment of their input gradients. We show theoretically and empirically that this induces the learning of more complex predictive patterns. OOD generalization fundamentally requires information beyond i.i.d. examples, such as multiple training environments, counterfactual examples, or other side information. Our approach shows that we can defer this requirement to an independent model selection stage. We obtain SOTA results in visual recognition on biased data and generalization across visual domains. The method - the first to evade the simplicity bias - highlights the need for a better understanding and control of inductive biases in deep learning.
翻译:通过SGD培训的神经网络最近被证明优先依赖线性预测特征,并且可以忽略复杂、同样预知的特征。这种简单偏差可以解释其分配缺乏稳健性的原因(OOD)。要学习的任务越复杂,就越有可能是统计文物(即选择偏向、虚假关联)比要学习的机制更简单。我们证明简单偏差是可以减轻的,OOOD一般化得到改进。我们训练了一套类似的模型,以不同的方式适应数据,使用对输入梯度调整的处罚。我们从理论上和经验上表明,这促使人们学习更复杂的预测模式。OOD一般化基本上需要多种培训环境、反事实实例或其他侧面信息等信息。我们的方法表明,我们可以将这一要求推迟到独立的模型选择阶段。我们从SOTA获得关于偏差数据以及视觉领域一般化的视觉识别结果。我们首先避免简单偏差的方法强调,需要更好地理解和控制深层次学习中的直觉偏差。