In spite of the high performance and reliability of deep learning algorithms in a wide range of everyday applications, many investigations tend to show that a lot of models exhibit biases, discriminating against specific subgroups of the population (e.g. gender, ethnicity). This urges the practitioner to develop fair systems with a uniform/comparable performance across sensitive groups. In this work, we investigate the gender bias of deep Face Recognition networks. In order to measure this bias, we introduce two new metrics, $\mathrm{BFAR}$ and $\mathrm{BFRR}$, that better reflect the inherent deployment needs of Face Recognition systems. Motivated by geometric considerations, we mitigate gender bias through a new post-processing methodology which transforms the deep embeddings of a pre-trained model to give more representation power to discriminated subgroups. It consists in training a shallow neural network by minimizing a Fair von Mises-Fisher loss whose hyperparameters account for the intra-class variance of each gender. Interestingly, we empirically observe that these hyperparameters are correlated with our fairness metrics. In fact, extensive numerical experiments on a variety of datasets show that a careful selection significantly reduces gender bias.
翻译:尽管在广泛的日常应用中深层次学习算法的性能和可靠性很高,但许多调查往往显示,许多模型表现出偏差,对特定人口分组(如性别、族裔)存在歧视。这促使从业人员制定公平制度,在敏感群体中实行统一/可比较的业绩;在这项工作中,我们调查深层认识网络的性别偏差;为了衡量这种偏差,我们引入了两个新标准,即$\mathrm{BFAR}和$\mathrm{BFRR},这更好地反映了面部识别系统的内在部署需要。受几何因素的驱使,我们通过新的后处理方法减少性别偏向,该方法将经过预先训练的模型的深度嵌入,赋予受歧视的分组更大的代表性权力。在这项工作中,我们通过尽量减少公平模式的米斯-菲斯-菲斯损失来培训浅层神经网络,而这种损失的高分量度是每个阶层内部差异的原因。有趣的是,我们从经验上看,这些超分量度计与我们的公平度度度度指标相关。事实上,我们通过新的后处理方法,我们通过一种新的处理方法来减轻性别选择。事实上,大量的数字实验显示各种数据的偏差。