Bias in classifiers is a severe issue of modern deep learning methods, especially for their application in safety- and security-critical areas. Often, the bias of a classifier is a direct consequence of a bias in the training dataset, frequently caused by the co-occurrence of relevant features and irrelevant ones. To mitigate this issue, we require learning algorithms that prevent the propagation of bias from the dataset into the classifier. We present a novel adversarial debiasing method, which addresses a feature that is spuriously connected to the labels of training images but statistically independent of the labels for test images. Thus, the automatic identification of relevant features during training is perturbed by irrelevant features. This is the case in a wide range of bias-related problems for many computer vision tasks, such as automatic skin cancer detection or driver assistance. We argue by a mathematical proof that our approach is superior to existing techniques for the abovementioned bias. Our experiments show that our approach performs better than state-of-the-art techniques on a well-known benchmark dataset with real-world images of cats and dogs.
翻译:分类法中的比亚语是现代深层次学习方法的一个严重问题,特别是在安全和保安关键领域应用的方法。通常,分类法的偏向是培训数据集中偏差的直接后果,这种偏差往往是相关特征和不相干特征共同出现造成的。为了缓解这一问题,我们需要学习算法,防止数据集偏差传播到分类器中。我们提出了一个新的对抗偏差方法,它处理一个与培训图像标签有虚假联系但统计上独立于测试图像标签的特征。因此,培训期间自动识别相关特征被不相关的特征所渗透。许多计算机视觉任务中与偏差有关的广泛问题就是这种情况,例如自动皮肤癌检测或司机协助。我们通过数学证明认为,我们的方法优于上述偏差的现有技术。我们的实验表明,我们的方法比在与猫和狗的真实世界图像相熟知的基准数据集上采用的最新技术要好。