Dataset biases are notoriously detrimental to model robustness and generalization. The identify-emphasize paradigm appears to be effective in dealing with unknown biases. However, we discover that it is still plagued by two challenges: A, the quality of the identified bias-conflicting samples is far from satisfactory; B, the emphasizing strategies only produce suboptimal performance. In this paper, for challenge A, we propose an effective bias-conflicting scoring method (ECS) to boost the identification accuracy, along with two practical strategies -- peer-picking and epoch-ensemble. For challenge B, we point out that the gradient contribution statistics can be a reliable indicator to inspect whether the optimization is dominated by bias-aligned samples. Then, we propose gradient alignment (GA), which employs gradient statistics to balance the contributions of the mined bias-aligned and bias-conflicting samples dynamically throughout the learning process, forcing models to leverage intrinsic features to make fair decisions. Furthermore, we incorporate self-supervised (SS) pretext tasks into training, which enable models to exploit richer features rather than the simple shortcuts, resulting in more robust models. Experiments are conducted on multiple datasets in various settings, demonstrating that the proposed solution can mitigate the impact of unknown biases and achieve state-of-the-art performance.
翻译:数据集偏差明显不利于模型的稳健性和概括性。识别强调范式似乎有效处理未知偏差。然而,我们发现它仍然受到两个挑战的困扰:A,所查明的偏见冲突抽样的质量远非令人满意;B,强调战略只产生亚优性表现。在本文中,为了挑战A,我们建议一种有效的偏见冲突评分方法(ECS)来提高识别准确性,同时采用两种实用战略 -- -- 同侪挑选和简单拼凑。关于挑战B,我们指出梯度贡献统计数据可以是一个可靠的指标,用来检查优化是否由偏差抽样所主导。然后,我们建议梯度调整(GA),在学习过程中,采用梯度统计来动态平衡被开采的偏差和偏见冲突抽样的贡献,迫使模型利用内在特征作出公平决定。此外,我们将自我监督(SS)的托辞任务纳入培训,使模型能够利用较富裕的特征,而不是简单的捷径,从而形成更健全的模型。我们建议对多种数据偏差的偏差性进行了实验,可以减少各种环境中的偏差性分析。