The widespread use of Artificial Intelligence (AI) in consequential domains, such as healthcare and parole decision-making systems, has drawn intense scrutiny on the fairness of these methods. However, ensuring fairness is often insufficient as the rationale for a contentious decision needs to be audited, understood, and defended. We propose that the attention mechanism can be used to ensure fair outcomes while simultaneously providing feature attributions to account for how a decision was made. Toward this goal, we design an attention-based model that can be leveraged as an attribution framework. It can identify features responsible for both performance and fairness of the model through attention interventions and attention weight manipulation. Using this attribution framework, we then design a post-processing bias mitigation strategy and compare it with a suite of baselines. We demonstrate the versatility of our approach by conducting experiments on two distinct data types, tabular and textual.
翻译:在保健与假释决策系统等相应领域广泛使用人工智能(AI),引起了对这些方法的公正性的严格审查,然而,确保公平性往往不够充分,因为争议性决定的理由需要审计、理解和辩护。我们提议,可利用关注机制确保公平结果,同时提供特征属性,说明如何作出决定。为实现这一目标,我们设计了一种以关注为基础的模式,可以作为归属框架加以利用。它可以通过关注干预和关注权重的操纵,确定模式业绩和公正性的负责特征。然后,我们利用这一归属框架,设计一个处理后减少偏见战略,并将其与一系列基线进行比较。我们通过对两种不同的数据类型,即表格和文字进行实验,展示了我们做法的多功能性。