With the proliferation of algorithmic decision-making, increased scrutiny has been placed on these systems. This paper explores the relationship between the quality of the training data and the overall fairness of the models trained with such data in the context of supervised classification. We measure key fairness metrics across a range of algorithms over multiple image classification datasets that have a varying level of noise in both the labels and the training data itself. We describe noise in the labels as inaccuracies in the labelling of the data in the training set and noise in the data as distortions in the data, also in the training set. By adding noise to the original datasets, we can explore the relationship between the quality of the training data and the fairness of the output of the models trained on that data.
翻译:随着算法决策的 proliferate,这些系统受到了越来越多的审视。本文探讨了在监督分类的上下文中训练数据质量与使用该等数据训练的模型整体公平性之间的关系。我们使用多个图像分类数据集,这些数据集在标签和训练数据本身的噪声水平都有所不同,对一系列算法进行测量以确定关键的公平度量指标。我们将标签中的噪声描述为训练集中数据标签的精度不准确,将数据中的噪声描述为训练集中数据扭曲。通过向原始数据集添加噪声,我们可以探讨训练数据质量与使用该等数据训练的模型的公平性之间的关系。