Recent studies showed that datasets used in fairness-aware machine learning for multiple protected attributes (referred to as multi-discrimination hereafter) are often imbalanced. The class-imbalance problem is more severe for the often underrepresented protected group (e.g. female, non-white, etc.) in the critical minority class. Still, existing methods focus only on the overall error-discrimination trade-off, ignoring the imbalance problem, thus amplifying the prevalent bias in the minority classes. Therefore, solutions are needed to solve the combined problem of multi-discrimination and class-imbalance. To this end, we introduce a new fairness measure, Multi-Max Mistreatment (MMM), which considers both (multi-attribute) protected group and class membership of instances to measure discrimination. To solve the combined problem, we propose a boosting approach that incorporates MMM-costs in the distribution update and post-training selects the optimal trade-off among accurate, balanced, and fair solutions. The experimental results show the superiority of our approach against state-of-the-art methods in producing the best balanced performance across groups and classes and the best accuracy for the protected groups in the minority class.
翻译:最近的研究显示,公平意识机器学习多重受保护属性(以下简称多重歧视)时使用的数据集往往不平衡,对少数群体中代表性往往不足的群体(如女性、非白人等)而言,等级不平衡问题更为严重,但现有方法仅侧重于总体错误-歧视权衡,忽视不平衡问题,从而扩大少数群体中普遍存在的偏见,因此,需要解决方案来解决多重歧视和阶级平衡的混合问题。为此,我们引入了新的公平措施,即多Max虐待(MMM),既考虑(多分配制)受保护群体成员,也考虑衡量歧视的类别成员。为解决这一合并问题,我们提议采取一种促进办法,将MMM-成本纳入分配更新和后培训,在准确、平衡和公平的解决办法中选择最佳的权衡。实验结果显示,我们的方法优于最平衡的跨群体和阶层业绩,以及保护群体的最佳准确性。