The increasing use of Machine Learning (ML) software can lead to unfair and unethical decisions, thus fairness bugs in software are becoming a growing concern. Addressing these fairness bugs often involves sacrificing ML performance, such as accuracy. To address this issue, we present a novel counterfactual approach that uses counterfactual thinking to tackle the root causes of bias in ML software. In addition, our approach combines models optimized for both performance and fairness, resulting in an optimal solution in both aspects. We conducted a thorough evaluation of our approach on 10 benchmark tasks using a combination of 5 performance metrics, 3 fairness metrics, and 15 measurement scenarios, all applied to 8 real-world datasets. The conducted extensive evaluations show that the proposed method significantly improves the fairness of ML software while maintaining competitive performance, outperforming state-of-the-art solutions in 84.6% of overall cases based on a recent benchmarking tool.
翻译:越来越多地使用机器学习软件可能导致不公平和不道德的决定,因此软件中的公平错误正在成为一个日益严重的问题。解决这些公平错误往往涉及牺牲ML性能,例如准确性。为了解决这一问题,我们提出了一个新的反事实方法,利用反事实思维解决ML软件中偏见的根源。此外,我们的方法结合了既有利于业绩又有利于公平两种方面的最佳模式,从而在两个方面都达成最佳解决办法。我们利用5个业绩计量、3个公平度和15个计量假设组合,对10项基准任务的方法进行了彻底评估,所有这些都适用于8个真实世界数据集。进行的广泛评估表明,拟议方法在保持竞争性业绩的同时,大大改善了ML软件的公平性,在84.6%的总体案件中,根据最近的基准工具,业绩优于最先进的解决办法。