Machine learning models are trained to minimize the mean loss for a single metric, and thus typically do not consider fairness and robustness. Neglecting such metrics in training can make these models prone to fairness violations when training data are imbalanced or test distributions differ. This work introduces Fairness Optimized Reweighting via Meta-Learning (FORML), a training algorithm that balances fairness and robustness with accuracy by jointly learning training sample weights and neural network parameters. The approach increases model fairness by learning to balance the contributions from both over- and under-represented sub-groups through dynamic reweighting of the data learned from a user-specified held-out set representative of the distribution under which fairness is desired. FORML improves equality of opportunity fairness criteria on image classification tasks, reduces bias of corrupted labels, and facilitates building more fair datasets via data condensation. These improvements are achieved without pre-processing data or post-processing model outputs, without learning an additional weighting function, without changing model architecture, and while maintaining accuracy on the original predictive metric.
翻译:培训机器学习模式是为了尽量减少单一指标的平均损失,因此通常不考虑公平性和稳健性。在培训数据不平衡或测试分布不同时,忽略这些指标会使这些模式容易受到公平违反。这项工作引入了公平性优化通过Meta-Linearch(FORML)重新加权(FORML)这一培训算法,这种算法通过联合学习培训样本重量和神经网络参数,既平衡公平性和稳健性,又平衡公平性和准确性。这一方法通过学习平衡代表性过大和不足的子群体的贡献,办法是通过动态地重新加权从用户指定的代表所选择的公平分布图象分配图集中获取的数据。FORMLML在图像分类任务方面改进机会平等标准,减少腐败标签的偏差,并通过数据凝固促进建立更公平的数据集。这些改进是在不学习额外加权功能,不改变模型结构,同时保持原始预测指标的准确性的情况下实现的。