In fair classification, it is common to train a model, and to compare and correct subgroup-specific error rates for disparities. However, even if a model's classification decisions satisfy a fairness metric, it is not necessarily the case that these decisions are equally confident. This becomes clear if we measure variance: We can fix everything in the learning process except the subset of training data, train multiple models, measure (dis)agreement in predictions for each test example, and interpret disagreement to mean that the learning process is more unstable with respect to its classification decision. Empirically, some decisions can in fact be so unstable that they are effectively arbitrary. To reduce this arbitrariness, we formalize a notion of self-consistency of a learning process, develop an ensembling algorithm that provably increases self-consistency, and empirically demonstrate its utility to often improve both fairness and accuracy. Further, our evaluation reveals a startling observation: Applying ensembling to common fair classification benchmarks can significantly reduce subgroup error rate disparities, without employing common pre-, in-, or post-processing fairness interventions. Taken together, our results indicate that variance, particularly on small datasets, can muddle the reliability of conclusions about fairness. One solution is to develop larger benchmark tasks. To this end, we release a toolkit that makes the Home Mortgage Disclosure Act datasets easily usable for future research.
翻译:在公平的分类中,常见的做法是训练一个模型,比较和纠正分小组的差错率。然而,即使模型的分类决定符合公平的衡量标准,这些决定也不一定具有同等的自信。如果我们衡量差异,这一点就变得十分清楚:我们可以解决学习过程中的一切问题,但培训数据分组、培训多种模型、每个测试示例预测中的(不一致)措施除外,并将分歧解释为学习过程与其分类决定相比更加不稳定。生动地说,有些决定事实上可能非常不稳定,以至于它们实际上具有任意性。为了减少这种任意性,我们正式确定学习过程的自我一致性概念,开发一种混合算法,可以增加自我一致性,并用经验证明它往往能提高公正和准确性。此外,我们的评估揭示出一种惊人的观察:采用共同的公平分类基准可以大大减少分组误差率差异,而不必使用共同的会前、内部或后处理公平性干预措施。加在一起,我们的结果表明,为了减少这种差异,特别是对于一个学习过程的自我一致性,我们制定一种混合的算法,能够增加内部数据的可靠性,从而得出一个更可靠的标准。