Machine Learning (ML) models are widely employed to drive many modern data systems. While they are undeniably powerful tools, ML models often demonstrate imbalanced performance and unfair behaviors. The root of this problem often lies in the fact that different subpopulations commonly display divergent trends: as a learning algorithm tries to identify trends in the data, it naturally favors the trends of the majority groups, leading to a model that performs poorly and unfairly for minority populations. Our goal is to improve the fairness and trustworthiness of ML models by applying only non-invasive interventions, i.e., without altering the data or the learning algorithm. We use a simple but key insight: the divergence of trends between different populations, and, consecutively, between a learned model and minority populations, is analogous to data drift, which indicates the poor conformance between parts of the data and the trained model. We explore two strategies (model-splitting and reweighing) to resolve this drift, aiming to improve the overall conformance of models to the underlying data. Both our methods introduce novel ways to employ the recently-proposed data profiling primitive of Conformance Constraints. Our experimental evaluation over 7 real-world datasets shows that both DifFair and ConFair improve the fairness of ML models. We demonstrate scenarios where DifFair has an edge, though ConFair has the greatest practical impact and outperforms other baselines. Moreover, as a model-agnostic technique, ConFair stays robust when used against different models than the ones on which the weights have been learned, which is not the case for other state of the art.
翻译:机器学习模型被广泛应用于推动现代数据系统。虽然它们无疑是强大的工具,但是机器学习模型经常表现出不平衡的性能和不公平的行为。问题的根源常常在于不同的亚群体常常显示不同的趋势:当学习算法尝试识别数据趋势时,自然地偏向多数群体的趋势,导致模型在少数群体中表现出不良的性能和不公平的行为。我们的目标是通过应用仅非侵入性干预(即不改变数据或学习算法)来提高机器学习模型的公平性和可信度。我们使用一个简单但关键的洞察:不同种群之间的趋势分歧,以及因此,学习模型和少数群体之间的分歧,类似于数据漂移,这表明数据中的某些部分与训练的模型之间的差距较大。我们探索了两种策略(模型分割和重新加权)来解决这种漂移,旨在提高模型对潜在数据的整体一致性。我们的两种方法都引入了最近提出的符合约束数据分析技术的创新应用。我们在7个真实世界的数据集上进行了实验评估,结果显示DifFair和ConFair都可以提高机器学习模型的公平性。我们展示了DifFair具有优势的情况,尽管ConFair具有最大的实际影响并且优于其他基线。此外,作为一个与模型无关的技术,ConFair在用于不同于学习权重的模型时保持稳健,这对其他最先进的技术不适用。