Despite being widely used, face recognition models suffer from bias: the probability of a false positive (incorrect face match) strongly depends on sensitive attributes such as the ethnicity of the face. As a result, these models can disproportionately and negatively impact minority groups, particularly when used by law enforcement. The majority of bias reduction methods have several drawbacks: they use an end-to-end retraining approach, may not be feasible due to privacy issues, and often reduce accuracy. An alternative approach is post-processing methods that build fairer decision classifiers using the features of pre-trained models. However, they still have drawbacks: they reduce accuracy (AGENDA, FTC), or require retuning for different false positive rates (FSN). In this work, we introduce the Fairness Calibration (FairCal) method, a post-training approach that: (i) increases model accuracy (improving the state-of-the-art), (ii) produces fairly-calibrated probabilities, (iii) significantly reduces the gap in the false positive rates, (iv) does not require knowledge of the sensitive attribute, and (v) does not require retraining, training an additional model, or retuning. We apply it to the task of Face Verification, and obtain state-of-the-art results with all the above advantages.
翻译:尽管广泛使用,但面对面的承认模式却存在偏见:假正(不正确的面部匹配)的概率在很大程度上取决于面部族裔等敏感属性。因此,这些模型可能不成比例地对少数群体产生消极影响,特别是在执法部门使用时。减少偏见的方法多数有几种缺点:它们采用端对端再培训方法,由于隐私问题,可能不可行,而且往往降低准确性。另一种办法是采用后处理方法,利用经过培训的模型的特点,建立更公平的决策分类,但是,它们仍然有缺陷:它们降低准确性(AGENDA、FTC),或要求对不同的假正率进行重新调整(FSN)。在这项工作中,我们采用了公平性校准(FairCal)方法,这是一种培训后方法,即:(一) 提高模型准确性(改进最新技术),(二) 产生相当有差异的概率,(三) 大大缩小假正率的差距,(四) 不需要了解敏感属性,(AGENDA, ETC) 和(VART) 要求采用更多的再培训结果。