Vision Transformer (ViT) has recently gained significant interest in solving computer vision (CV) problems due to its capability of extracting informative features and modeling long-range dependencies through the self-attention mechanism. To fully realize the advantages of ViT in real-world applications, recent works have explored the trustworthiness of ViT, including its robustness and explainability. However, another desiderata, fairness has not yet been adequately addressed in the literature. We establish that the existing fairness-aware algorithms (primarily designed for CNNs) do not perform well on ViT. This necessitates the need for developing our novel framework via Debiased Self-Attention (DSA). DSA is a fairness-through-blindness approach that enforces ViT to eliminate spurious features correlated with the sensitive attributes for bias mitigation. Notably, adversarial examples are leveraged to locate and mask the spurious features in the input image patches. In addition, DSA utilizes an attention weights alignment regularizer in the training objective to encourage learning informative features for target prediction. Importantly, our DSA framework leads to improved fairness guarantees over prior works on multiple prediction tasks without compromising target prediction performance
翻译:最近,由于能够通过自省机制提取信息特征和建模长期依赖性模型,人们最近对解决计算机视觉(CVV)问题产生了极大兴趣。为了充分实现VT在现实应用中的优势,最近的工作探索了VT的可信赖性,包括其稳健性和可解释性。然而,文献中尚未充分处理另一个贬低性、公平性的问题。我们确定,现有的公平觉悟算法(主要为CNNs设计)在VT上效果不佳。这需要通过DSA来开发我们的新颖框架。DSA是一种公平透视方法,强制VT消除与减少偏差的敏感属性相关的虚假特征。值得注意的是,利用了对抗性实例来查找和遮掩输入图像补补处中的虚假特征。此外,DSA在培训目标中利用了关注重量校准调节器鼓励学习目标预测的信息特征。重要的是,我们DSA框架在不牺牲多个预测目标之前的工作上改进了公平性保证。