Neural networks are prone to be biased towards spurious correlations between classes and latent attributes exhibited in a major portion of training data, which ruins their generalization capability. We propose a new method for training debiased classifiers with no spurious attribute label. The key idea is to employ a committee of classifiers as an auxiliary module that identifies bias-conflicting data, i.e., data without spurious correlation, and assigns large weights to them when training the main classifier. The committee is learned as a bootstrapped ensemble so that a majority of its classifiers are biased as well as being diverse, and intentionally fail to predict classes of bias-conflicting data accordingly. The consensus within the committee on prediction difficulty thus provides a reliable cue for identifying and weighting bias-conflicting data. Moreover, the committee is also trained with knowledge transferred from the main classifier so that it gradually becomes debiased along with the main classifier and emphasizes more difficult data as training progresses. On five real-world datasets, our method outperforms prior arts using no spurious attribute label like ours and even surpasses those relying on bias labels occasionally.
翻译:神经网络容易偏向于在大部分培训数据中显示的阶级和潜在属性之间虚假的关联,这破坏了它们的概括能力。我们提出一种新的方法来培训没有虚假属性标签的被贬低的分类者。关键的想法是使用一个分类者委员会作为辅助模块,确定偏见冲突数据,即没有虚假关联的数据,并在培训主要分类者时给它们分配大量权重。委员会作为一个累累的合谋学习,以便其分类者大多具有偏见和多样性,并故意不相应预测偏见冲突数据类别。预测委员会内的共识因此为识别和加权偏见冲突数据提供了可靠的提示。此外,委员会还接受主要分类者传授的知识培训,以便逐渐与主要分类者一道分化,强调培训进展中更难的数据。在五个真实世界数据集中,我们的方法超越了前科艺术,使用了没有虚假的属性标签,甚至超越了偶尔依赖偏见标签的人。