Algorithms and Machine Learning (ML) are increasingly affecting everyday life and several decision-making processes, where ML has an advantage due to scalability or superior performance. Fairness in such applications is crucial, where models should not discriminate their results based on race, gender, or other protected groups. This is especially crucial for models affecting very sensitive topics, like interview hiring or recidivism prediction. Fairness is not commonly studied for regression problems compared to binary classification problems; hence, we present a simple, yet effective method based on normalisation (FaiReg), which minimises the impact of unfairness in regression problems, especially due to labelling bias. We present a theoretical analysis of the method, in addition to an empirical comparison against two standard methods for fairness, namely data balancing and adversarial training. We also include a hybrid formulation (FaiRegH), merging the presented method with data balancing, in an attempt to face labelling and sample biases simultaneously. The experiments are conducted on the multimodal dataset First Impressions (FI) with various labels, namely personality prediction and interview screening score. The results show the superior performance of diminishing the effects of unfairness better than data balancing, also without deteriorating the performance of the original problem as much as adversarial training.
翻译:与二进制分类问题相比,对回归问题通常不进行公平研究;因此,我们提出了一个基于正常化(FaiReg)的简单而有效的方法(FaiReg),该方法最大限度地减少回归问题中的不公平影响,特别是由于贴标签偏差造成的不公问题的影响;我们对这种方法进行理论分析,除了与两种标准公平方法(即数据平衡和对抗性培训)进行实验性比较外,我们还对方法进行理论性分析;我们还包括一种混合法(FairegH),将提出的方法与数据平衡结合起来,以同时面对标签和抽样偏差;我们用多种标签(即个性预测和面试评分)对多式联运数据集进行了实验;结果显示,在减少不平等性影响方面表现优于数据平衡的初始性问题,同时没有恶化。