Neural NLP models tend to rely on spurious correlations between labels and input features to perform their tasks. Minority examples, i.e., examples that contradict the spurious correlations present in the majority of data points, have been shown to increase the out-of-distribution generalization of pre-trained language models. In this paper, we first propose using example forgetting to find minority examples without prior knowledge of the spurious correlations present in the dataset. Forgettable examples are instances either learned and then forgotten during training or never learned. We empirically show how these examples are related to minorities in our training sets. Then, we introduce a new approach to robustify models by fine-tuning our models twice, first on the full training data and second on the minorities only. We obtain substantial improvements in out-of-distribution generalization when applying our approach to the MNLI, QQP, and FEVER datasets.
翻译:少数群体的例子,即与大多数数据点存在的虚假关联相矛盾的例子,被证明可提高预先培训的语言模式在分配方面的普遍化。在本文中,我们首先提议在事先不知道数据集存在的虚假关联的情况下,以实例忘记寻找少数群体的例子。可被遗忘的例子有的在培训过程中学到的、然后被遗忘的或从未学过的例子。我们从经验上展示了这些例子与我们的培训组合中的少数群体的关系。然后,我们引入了一种新的方法,通过对模型进行两次微调,首先对全部培训数据进行微调,其次仅对少数群体进行微调,强化模型。我们在对MNLI、QP和Fever数据集应用我们的方法时,在分配以外的概括化方面有了实质性的改进。