As transformer architectures become increasingly prevalent in computer vision, it is critical to understand their fairness implications. We perform the first study of the fairness of transformers applied to computer vision and benchmark several bias mitigation approaches from prior work. We visualize the feature space of the transformer self-attention modules and discover that a significant portion of the bias is encoded in the query matrix. With this knowledge, we propose TADeT, a targeted alignment strategy for debiasing transformers that aims to discover and remove bias primarily from query matrix features. We measure performance using Balanced Accuracy and Standard Accuracy, and fairness using Equalized Odds and Balanced Accuracy Difference. TADeT consistently leads to improved fairness over prior work on multiple attribute prediction tasks on the CelebA dataset, without compromising performance.
翻译:随着变压器结构在计算机愿景中日益普遍,了解其公平影响至关重要。我们首次研究了用于计算机愿景的变压器的公平性,并将先前工作中的若干减少偏差的方法基准化。我们设想了变压器自我注意模块的特征空间,发现很大一部分偏差在查询矩阵中被编码。有了这一知识,我们建议TADeT(TADET),这是针对贬低变压器的定向调整战略,主要目的是发现并消除查询矩阵特征中的偏差。我们用平衡准确性和标准准确性来衡量业绩,用平衡奇数和平衡准确性差异来衡量公平性。TADET(TADET)一贯地在不损害业绩的情况下,提高以前在CelebA数据集多个属性预测任务上的工作的公正性。