Human lives are increasingly being affected by the outcomes of automated decision-making systems and it is essential for the latter to be, not only accurate, but also fair. The literature of algorithmic fairness has grown considerably over the last decade, where most of the approaches are evaluated under the strong assumption that the train and test samples are independently and identically drawn from the same underlying distribution. However, in practice, dissimilarity between the training and deployment environments exists, which compromises the performance of the decision-making algorithm as well as its fairness guarantees in the deployment data. There is an emergent research line that studies how to preserve fairness guarantees when the data generating processes differ between the source (train) and target (test) domains, which is growing remarkably. With this survey, we aim to provide a wide and unifying overview on the topic. For such purpose, we propose a taxonomy of the existing approaches for fair classification under distribution shift, highlight benchmarking alternatives, point out the relation with other similar research fields and eventually, identify future venues of research.
翻译:人类生活日益受到自动化决策系统结果的影响,后者不仅必须准确,而且必须公平。算法公正文献在过去十年中有了相当大的发展,在过去十年中,大多数方法都根据一个强有力的假设进行评估,即火车和测试样品是独立和完全从同一基本分布中提取的。然而,在实践中,培训和部署环境之间存在差异,这损害了决策算法的性能及其在部署数据中的公平保障。有一个新兴的研究线,研究如何在数据生成过程在来源(火车)和目标(试验)领域之间存在差异时维护公平保障,这种差异正在显著增加。我们通过这项调查,我们力求就这一主题提供一个广泛和统一的概览。为此,我们提议对分配转移下的公平分类的现有方法进行分类,突出基准备选方案,指出与其他类似的研究领域的关系,并最终确定未来的研究地点。