The remarkable performance of deep learning models and their applications in consequential domains (e.g., facial recognition) introduces important challenges at the intersection of equity and security. Fairness and robustness are two desired notions often required in learning models. Fairness ensures that models do not disproportionately harm (or benefit) some groups over others, while robustness measures the models' resilience against small input perturbations. This paper shows the existence of a dichotomy between fairness and robustness, and analyzes when achieving fairness decreases the model robustness to adversarial samples. The reported analysis sheds light on the factors causing such contrasting behavior, suggesting that distance to the decision boundary across groups as a key explainer for this behavior. Extensive experiments on non-linear models and different architectures validate the theoretical findings in multiple vision domains. Finally, the paper proposes a simple, yet effective, solution to construct models achieving good tradeoffs between fairness and robustness.
翻译:深层次学习模式及其在相应领域应用的显著表现(如面部识别)在公平和安全交叉方面提出了重大挑战。公平性和稳健性是学习模式经常需要的两个理想概念。公平性确保模型不会对一些群体造成不成比例的伤害(或好处),而稳健性则衡量模型在小型投入干扰方面的抗御力。本文表明公平性和稳健性之间存在二分法,并在实现公平性时分析模型对对抗性样本的稳健性。所报告的分析揭示了造成这种反差行为的因素,表明各群体之间与决定界限的距离是这一行为的关键解释者。非线性模型和不同结构的广泛实验验证了多愿景领域的理论结论。最后,本文件提出了一个简单而有效的解决方案,用以构建在公平和稳健之间实现良好权衡的模型。