Many recent works indicate that the deep neural networks tend to take dataset biases as shortcuts to make decision, rather than understand the tasks, which results in failures on the real-world applications. In this work, we focus on the spurious correlation between feature and label, which derive from the biased data distribution in the training data, and analyze it concretely. In particular, we define the word highly co-occurring with a specific label as biased word, and the example containing biased word as biased example. Our analysis reveals that the biased examples with spurious correlations are easier for models to learn, and when predicting, the biased words make significantly higher contributions to models' predictions than other words, and the models tend to assign the labels over-relying on the spurious correlation between words and labels. To mitigate the model's over-reliance on the shortcut, we propose a training strategy Less-Learn-Shortcut (LLS): we quantify the biased degree of the biased examples, and down-weight them with the biased degree. Experimental results on QM and NLI tasks show that the models improve the performances both on in-domain and adversarial data (1.57% on DuQM and 2.12% on HANS) with our LLS.
翻译:许多最近的工作都表明,深心神经网络倾向于将数据集偏差作为决策的捷径,而不是理解任务,从而导致真实世界应用的失败。在这项工作中,我们侧重于特征和标签之间的虚假关联,这种关联源自培训数据中的数据分布偏差,并具体分析。特别是,我们将与特定标签高度相联的词定义为有偏见的单词,并将含有偏颇词的例子定义为有偏见的例子。我们的分析表明,带有虚假相关性的偏差例子对于模型的学习比较容易,当预测时,偏差词对模型预测的贡献远远大于其他词,而模型往往把标签过多地放在词和标签之间的虚假关联上。为了减轻模型过分依赖捷径的情况,我们提议了一个培训战略“少利恩捷道”:我们量化偏差例子的偏差程度,用偏差程度来降低它们。QM和NLIL任务实验结果显示,模型在DAM数据上和DAM数据上改进了内部和DABM数据上的绩效(1%)。