The fairness in machine learning is getting increasing attention, as its applications in different fields continue to expand and diversify. To mitigate the discriminated model behaviors between different demographic groups, we introduce a novel post-processing method to optimize over multiple fairness constraints through group-aware threshold adaptation. We propose to learn adaptive classification thresholds for each demographic group by optimizing the confusion matrix estimated from the probability distribution of a classification model output. As we only need an estimated probability distribution of model output instead of the classification model structure, our post-processing model can be applied to a wide range of classification models and improve fairness in a model-agnostic manner and ensure privacy. This even allows us to post-process existing fairness methods to further improve the trade-off between accuracy and fairness. Moreover, our model has low computational cost. We provide rigorous theoretical analysis on the convergence of our optimization algorithm and the trade-off between accuracy and fairness of our method. Our method theoretically enables a better upper bound in near optimality than existing method under same condition. Experimental results demonstrate that our method outperforms state-of-the-art methods and obtains the result that is closest to the theoretical accuracy-fairness trade-off boundary.
翻译:机器学习的公平性正在日益引起人们的注意,因为其在不同领域的应用继续扩大和多样化。为了减少不同人口群体之间的歧视模式行为,我们采用了一种新的后处理方法,通过群体认知阈值的适应性调整,优化多重公平限制;我们提议通过优化根据分类模型产出的概率分布估计出的混乱矩阵,为每个人口群体学习适应性分类阈值;由于我们只需要估计模型产出的概率分布,而不是分类模型结构,我们的后处理模型可以应用到广泛的分类模型中,并改进以模型-不可知的方式的公平性,并确保隐私。这甚至使我们能够处理现有的公平方法,以进一步改进准确性和公正性之间的权衡。此外,我们的模型的计算成本较低。我们从理论上对我们优化算法的趋同和我们方法的准确性和公平性之间的权衡进行了严格的理论分析。我们的方法理论上可以比在同样条件下的现有方法更接近最优化的上限。实验结果表明,我们的方法超越了最先进的标准,并获得了最接近理论准确性公平贸易界限的结果。