As equality issues in the use of face recognition have garnered a lot of attention lately, greater efforts have been made to debiased deep learning models to improve fairness to minorities. However, there is still no clear definition nor sufficient analysis for bias assessment metrics. We propose an information-theoretic, independent bias assessment metric to identify degree of bias against protected demographic attributes from learned representations of pretrained facial recognition systems. Our metric differs from other methods that rely on classification accuracy or examine the differences between ground truth and predicted labels of protected attributes predicted using a shallow network. Also, we argue, theoretically and experimentally, that logits-level loss is not adequate to explain bias since predictors based on neural networks will always find correlations. Further, we present a synthetic dataset that mitigates the issue of insufficient samples in certain cohorts. Lastly, we establish a benchmark metric by presenting advantages in clear discrimination and small variation comparing with other metrics, and evaluate the performance of different debiased models with the proposed metric.
翻译:由于使用面部识别方面的平等问题最近引起人们的极大关注,我们已作出更大努力,贬低深层次学习模式,以提高对少数群体的公平性;然而,对于偏向评估指标,尚没有明确的定义和充分分析;我们建议采用信息理论独立的偏向评估尺度,从事先经过训练的面部识别系统的学术表现中找出对受保护人口特征的偏向程度;我们的衡量尺度不同于依赖分类精确度或审查地面真相与利用浅线网预测的保护属性的预测标志之间的差异的其他方法;此外,我们主张,在理论上和实验上,对记录级别的损失不足以解释偏向,因为基于神经网络的预测器总是会发现关联性。此外,我们提出了一个综合数据集,以缓解某些组群中样本不足的问题。最后,我们确立了一个基准衡量尺度,在与其他指标相比明显歧视和小变异方面提供了优势,并评估了与拟议指标不同贬低模型的性模型的性能。