Deep learning models generally learn the biases present in the training data. Researchers have proposed several approaches to mitigate such biases and make the model fair. Bias mitigation techniques assume that a sufficiently large number of training examples are present. However, we observe that if the training data is limited, then the effectiveness of bias mitigation methods is severely degraded. In this paper, we propose a novel approach to address this problem. Specifically, we adapt self-supervision and self-distillation to reduce the impact of biases on the model in this setting. Self-supervision and self-distillation are not used for bias mitigation. However, through this work, we demonstrate for the first time that these techniques are very effective in bias mitigation. We empirically show that our approach can significantly reduce the biases learned by the model. Further, we experimentally demonstrate that our approach is complementary to other bias mitigation strategies. Our approach significantly improves their performance and further reduces the model biases in the limited data regime. Specifically, on the L-CIFAR-10S skewed dataset, our approach significantly reduces the bias score of the baseline model by 78.22% and outperforms it in terms of accuracy by a significant absolute margin of 8.89%. It also significantly reduces the bias score for the state-of-the-art domain independent bias mitigation method by 59.26% and improves its performance by a significant absolute margin of 7.08%.
翻译:深层学习模式通常会了解培训数据中存在的偏见。研究人员提出了几种减少这种偏见并使模型公平化的办法。Bias减缓技术假定存在足够多的培训实例。然而,我们观察到,如果培训数据有限,那么减少偏见方法的有效性就会严重退化。我们在本文件中提出了解决这一问题的新颖办法。具体地说,我们调整了自我监督与自我蒸馏方法,以减少偏见对模型在这个环境中的影响。自我监督与自我蒸馏方法不被用于减少偏见。然而,通过这项工作,我们第一次证明这些技术在减少偏见方面非常有效。我们从经验上表明,我们的方法可以大大减少从模型中学到的偏见。此外,我们实验性地表明,我们的方法是对其他减少偏见战略的补充。我们的方法大大改进了它们的业绩,并进一步减少了有限数据制度中的模型偏差。具体地说,在L-CIFAR-10S扭曲数据集中,我们的方法大大降低了基线模型的偏差分数,减少了78.22%,在减少这种偏差方面,我们第一次表明这些技术在减少偏见方面非常有效。我们实验表明,我们的方法可以大大地减少从模型中学到的偏差差值。