Several proposals have been put forward in recent years for improving out-of-distribution (OOD) performance through mitigating dataset biases. A popular workaround is to train a robust model by re-weighting training examples based on a secondary biased model. Here, the underlying assumption is that the biased model resorts to shortcut features. Hence, those training examples that are correctly predicted by the biased model are flagged as being biased and are down-weighted during the training of the main model. However, assessing the importance of an instance merely based on the predictions of the biased model may be too naive. It is possible that the prediction of the main model can be derived from another decision-making process that is distinct from the behavior of the biased model. To circumvent this, we introduce a fine-tuning strategy that incorporates the similarity between the main and biased model attribution scores in a Product of Experts (PoE) loss function to further improve OOD performance. With experiments conducted on natural language inference and fact verification benchmarks, we show that our method improves OOD results while maintaining in-distribution (ID) performance.
翻译:近些年来,为通过减少数据集偏差来改善分配(OOD)不分配(OOOD)的绩效,提出了几项建议; 大众变通办法是通过根据次要偏差模式对培训范例进行重新加权,来训练一个强有力的模型; 这里的基本假设是,偏差模式诉诸于捷径特征; 因此,偏差模式正确预测的那些培训范例被贴上偏见的标签,在培训主要模型期间被降级加权; 然而,仅仅根据对偏差模式的预测来评估实例的重要性可能过于天真; 对主要模型的预测可能来自与偏差模式不同的另一个决策进程; 为了绕过这一假设,我们引入了一种微调战略,将专家产品(PoE)损失功能中的主要和偏差模式分配分数的相似性纳入其中,以进一步提高OODD的绩效; 在对自然语言推断和事实核实基准进行实验后,我们发现我们的方法在保持分配(ID)业绩的同时改进OD的结果。