As machine learning (ML) algorithms are increasingly used in high-stakes applications, concerns have arisen that they may be biased against certain social groups. Although many approaches have been proposed to make ML models fair, they typically rely on the assumption that data distributions in training and deployment are identical. Unfortunately, this is commonly violated in practice and a model that is fair during training may lead to an unexpected outcome during its deployment. Although the problem of designing robust ML models under dataset shifts has been widely studied, most existing works focus only on the transfer of accuracy. In this paper, we study the transfer of both fairness and accuracy under domain generalization where the data at test time may be sampled from never-before-seen domains. We first develop theoretical bounds on the unfairness and expected loss at deployment, and then derive sufficient conditions under which fairness and accuracy can be perfectly transferred via invariant representation learning. Guided by this, we design a learning algorithm such that fair ML models learned with training data still have high fairness and accuracy when deployment environments change. Experiments on real-world data validate the proposed algorithm. Model implementation is available at https://github.com/pth1993/FATDM.
翻译:由于机器学习(ML)算法越来越多地用于高镜头应用,人们担心这些算法可能对某些社会群体有偏向,虽然提出许多办法使ML模型公平,但它们通常依赖的假设是,培训和部署中的数据分布是相同的;不幸的是,这种做法在实践上普遍违反,在培训期间公平的模式可能会在部署过程中产生出意想不到的结果。虽然在数据集变化下设计稳健的ML模型的问题已经得到广泛研究,但大多数现有工作只侧重于准确性转移。在本文中,我们研究在域通用下,公平性和准确性的转移,在测试时的数据可以从从未见的领域取样出来。我们首先就部署时的不公平和预期损失制定了理论界限,然后得出足够的条件,使公平和准确性能够完全通过动态代表学习来转移。我们以此为指导设计了一种学习算法,使利用培训数据学习的公平 ML模型在部署环境变化时仍然具有高度的公正性和准确性。关于真实世界数据实验验证拟议算法。模型的实施可以在 https://github.M/comth/1993。