As a technology ML is oblivious to societal good or bad, and thus, the field of fair machine learning has stepped up to propose multiple mathematical definitions, algorithms, and systems to ensure different notions of fairness in ML applications. Given the multitude of propositions, it has become imperative to formally verify the fairness metrics satisfied by different algorithms on different datasets. In this paper, we propose a stochastic satisfiability (SSAT) framework, Justicia, that formally verifies different fairness measures of supervised learning algorithms with respect to the underlying data distribution. We instantiate Justicia on multiple classification and bias mitigation algorithms, and datasets to verify different fairness metrics, such as disparate impact, statistical parity, and equalized odds. Justicia is scalable, accurate, and operates on non-Boolean and compound sensitive attributes unlike existing distribution-based verifiers, such as FairSquare and VeriFair. Being distribution-based by design, Justicia is more robust than the verifiers, such as AIF360, that operate on specific test samples. We also theoretically bound the finite-sample error of the verified fairness measure.
翻译:由于技术ML忽视了社会上的好坏,因此,公平机器学习领域已加速提出多种数学定义、算法和系统,以确保多数学定义、算法和系统,以确保多数学应用中的公平概念。鉴于各种主张,现在必须正式核实不同数据集的不同算法所满足的公平度量标准。在本文件中,我们提议了一个Stochistic相对性(SSAT)框架,即“Justicia”,正式核查与基本数据分布有关的监督学习算法的不同公平度度量。我们即时在多种分类和减少偏差的算法和数据集方面提出正义,以核实不同的公平度量度标准,如不同的影响、统计均等和均等的几率。“Justicia”是可伸缩的,准确的,并且以非Boolean和复合的敏感属性运作,不同于现有的基于分配的校验标准,如FairSquare和VeriFaiir。根据设计进行分配,Justicia比核查者更强大,如AIF360,我们理论上也约束了具体测试样品所核查的公平度度度度误差。