Machine Learning (ML) models are being used in all facets of today's society to make high stake decisions like bail granting or credit lending, with very minimal regulations. Such systems are extremely vulnerable to both propagating and amplifying social biases, and have therefore been subject to growing research interest. One of the main issues with conventional fairness metrics is their narrow definitions which hide the complete extent of the bias by focusing primarily on positive and/or negative outcomes, whilst not paying attention to the overall distributional shape. Moreover, these metrics are often contradictory to each other, are severely restrained by the contextual and legal landscape of the problem, have technical constraints like poor support for continuous outputs, the requirement of class labels, and are not explainable. In this paper, we present Quantile Demographic Drift, which addresses the shortcomings mentioned above. This metric can also be used to measure intra-group privilege. It is easily interpretable via existing attribution techniques, and also extends naturally to individual fairness via the principle of like-for-like comparison. We make this new fairness score the basis of a new system that is designed to detect bias in production ML models without the need for labels. We call the system FairCanary because of its capability to detect bias in a live deployed model and narrow down the alert to the responsible set of features, like the proverbial canary in a coal mine.
翻译:当今社会各个方面都在使用机器学习模式,以极低的监管条例,在保释或信贷贷款等最起码的监管下,在当今社会所有方面作出重大利益决策。这些制度极易受到社会偏见的传播和扩张,因此引起越来越多的研究兴趣。传统公平衡量标准的主要问题之一是其狭隘的定义掩盖了偏见的全部范围,主要侧重于正面和(或)负面结果,同时不注意整体分配形式。此外,这些衡量标准往往相互矛盾,受到问题的背景和法律背景的严重制约,技术制约,如对连续产出的支持不力、阶级标签要求等技术限制,无法解释。在本文件中,我们介绍解决上述缺陷的量化人口动态指标。该指标也可以用来衡量群体内部特权的完整程度,通过现有归属技术很容易解释,并通过类似比较原则自然地延伸到个人公平。我们让这种新的公平分数制度的基础成为了设计用来检测生产模型中的偏差,如对阶级标签的要求,并且无法解释。我们在本文中提出“量化人口动态”Drift,因为我们可以在不需要负责性标签的情况下,用一个狭隘的标志来测量。