Prior work has shown that Visual Recognition datasets frequently under-represent sensitive groups (\eg Female) within a category (\eg Programmers). This dataset bias can lead to models that learn spurious correlations between class labels and sensitive attributes such as age, gender, or race. Most of the recent methods that address this problem require significant architectural changes or expensive hyper-parameter tuning. Alternatively, data re-sampling baselines from the class imbalance literature (\eg Undersampling, Upweighting), which can often be implemented in a single line of code and often have no hyperparameters, offer a cheaper and more efficient solution. However, we found that some of these baselines were missing from recent bias mitigation benchmarks. In this paper, we show that these simple methods are strikingly competitive with state-of-the-art bias mitigation methods on many datasets. Furthermore, we improve these methods by introducing a new class conditioned sampling method: Bias Mimicking. In cases where the baseline dataset re-sampling methods do not perform well, Bias Mimicking effectively bridges the performance gap and improves the total averaged accuracy of under-represented subgroups by over $3\%$ compared to prior work.
翻译:先前的工作表明,视觉识别数据集经常在一个类别(\eg Programmers)内为敏感群体( eg women) 提供较低比例的数据。这种数据集偏差可能导致一些模型,这些模型可以了解等级标签与年龄、性别或种族等敏感属性之间的虚假关联。大多数解决这一问题的最近方法都需要重大的建筑变化或昂贵的超参数调整。或者,从类别不平衡文献(\\eg surrupsamping, 提高加权)中重新抽取数据基线,这些基线数据往往可以在单行代码中执行,而且往往没有超分光度,提供更便宜、更有效率的解决办法。然而,我们发现有些基线在近期的减少偏差基准中缺失。我们在本文件中表明,这些简单方法与许多数据集中最先进的减少偏差的方法相比,具有惊人的竞争力。此外,我们通过采用新的有条件的分类抽样方法改进这些方法: Bias Mimicking。如果基准数据集重新取样方法不完善,那么Bias Mimicking 有效地弥补了业绩差距,并改进了先前分组工作的总平均值。