Accuracy and individual fairness are both crucial for trustworthy machine learning, but these two aspects are often incompatible with each other so that enhancing one aspect may sacrifice the other inevitably with side effects of true bias or false fairness. We propose in this paper a new fairness criterion, accurate fairness, to align individual fairness with accuracy. Informally, it requires the treatments of an individual and the individual's similar counterparts to conform to a uniform target, i.e., the ground truth of the individual. We prove that accurate fairness also implies typical group fairness criteria over a union of similar sub-populations. We then present a Siamese fairness in-processing approach to minimize the accuracy and fairness losses of a machine learning model under the accurate fairness constraints. To the best of our knowledge, this is the first time that a Siamese approach is adapted for bias mitigation. We also propose fairness confusion matrix-based metrics, fair-precision, fair-recall, and fair-F1 score, to quantify a trade-off between accuracy and individual fairness. Comparative case studies with popular fairness datasets show that our Siamese fairness approach can achieve on average 1.02%-8.78% higher individual fairness (in terms of fairness through awareness) and 8.38%-13.69% higher accuracy, as well as 10.09%-20.57% higher true fair rate, and 5.43%-10.01% higher fair-F1 score, than the state-of-the-art bias mitigation techniques. This demonstrates that our Siamese fairness approach can indeed improve individual fairness without trading accuracy. Finally, the accurate fairness criterion and Siamese fairness approach are applied to mitigate the possible service discrimination with a real Ctrip dataset, by on average fairly serving 112.33% more customers (specifically, 81.29% more customers in an accurately fair way) than baseline models.
翻译:准确性和个人公平性对于值得信赖的机器学习都至关重要,但这两个公平性往往不相容,因此,加强一个方面可能会不可避免地牺牲另一个方面,而真正的偏差或虚假的公平性会产生副作用。我们在本文件中提出了一个新的公平性标准,即准确性,以使个人公平性与准确性相一致。非正式地说,它要求个人和个人相似的对应方的待遇符合一个统一的目标,即个人的基本真理。我们证明,准确性也意味着对相似的子群体联盟的典型群体公平性标准。我们然后在处理中提出一个Siame的公平性方法,以在准确性制约下将机器学习模型的准确性和公正性损失降到最低。我们最了解的是,一个准确性标准,即准确性,即准确性,公平性,即准确性,即准确性,即准确性。我们用Siames的公平性方法在平均1.02-8-8中可以达到更高的标准。