There are several bias mitigators that can reduce algorithmic bias in machine learning models but, unfortunately, the effect of mitigators on fairness is often not stable when measured across different data splits. A popular approach to train more stable models is ensemble learning. Ensembles, such as bagging, boosting, voting, or stacking, have been successful at making predictive performance more stable. One might therefore ask whether we can combine the advantages of bias mitigators and ensembles? To explore this question, we first need bias mitigators and ensembles to work together. We built an open-source library enabling the modular composition of 10 mitigators, 4 ensembles, and their corresponding hyperparameters. Based on this library, we empirically explored the space of combinations on 13 datasets, including datasets commonly used in fairness literature plus datasets newly curated by our library. Furthermore, we distilled the results into a guidance diagram for practitioners. We hope this paper will contribute towards improving stability in bias mitigation.
翻译:但不幸的是,当通过不同的数据分割进行测量时,减轻因素对公平性的影响往往不稳定。 培养更稳定的模型的流行方法就是共同学习。 集成,如包装、推动、投票或堆叠,成功地使预测性能更加稳定。 因此,人们可能会问,我们是否可以将偏向缓解器和集合器的优点结合起来? 要探讨这一问题,我们首先需要偏见缓解器和集合器一起工作。 我们建立了一个开放源码图书馆,使10个缓解器、4个聚合器及其相应的超光谱计的模块组成成为可能。 在这个图书馆的基础上,我们从经验上探索了13个数据集的组合空间,包括公平文献中常用的数据集,加上我们图书馆新近整理的数据集。此外,我们把结果提炼成一个供开业者使用的指导图表。 我们希望这份文件将有助于改善偏见缓解的稳定性。