The fairness characteristic is a critical attribute of trusted AI systems. A plethora of research has proposed diverse methods for individual fairness testing. However, they are suffering from three major limitations, i.e., low efficiency, low effectiveness, and model-specificity. This work proposes ExpGA, an explanationguided fairness testing approach through a genetic algorithm (GA). ExpGA employs the explanation results generated by interpretable methods to collect high-quality initial seeds, which are prone to derive discriminatory samples by slightly modifying feature values. ExpGA then adopts GA to search discriminatory sample candidates by optimizing a fitness value. Benefiting from this combination of explanation results and GA, ExpGA is both efficient and effective to detect discriminatory individuals. Moreover, ExpGA only requires prediction probabilities of the tested model, resulting in a better generalization capability to various models. Experiments on multiple real-world benchmarks, including tabular and text datasets, show that ExpGA presents higher efficiency and effectiveness than four state-of-the-art approaches.
翻译:公平性是可信赖的AI系统的重要特征。 大量研究提出了个人公平性测试的多种方法,然而,它们受到三大限制,即低效率、低效率、低效力和模式特性。 这项工作提出了通过基因算法(GA)进行解释性公平性测试的ODGA(解释性指导性公平性测试方法)。 ExGA采用可解释性方法得出的解释性结果来收集高质量的初始种子,这些种子容易通过稍微修改特征值获得歧视性样本。 ExpGA随后通过GA(GA)来通过优化健康价值来搜索歧视性样本候选人。 从这一解释结果和GA(GA)的组合中得益,ExGA(ExGA)是发现歧视个人的高效率和有效性。 此外,ExGA(ExGA)仅要求预测测试模型的概率,从而提高各种模型的普及能力。关于多个真实世界基准的实验,包括表格和文本数据集,表明ExGA(ExGA)的效率和效力高于四种最先进的方法。