Selective regression allows abstention from prediction if the confidence to make an accurate prediction is not sufficient. In general, by allowing a reject option, one expects the performance of a regression model to increase at the cost of reducing coverage (i.e., by predicting on fewer samples). However, as we show, in some cases, the performance of a minority subgroup can decrease while we reduce the coverage, and thus selective regression can magnify disparities between different sensitive subgroups. Motivated by these disparities, we propose new fairness criteria for selective regression requiring the performance of every subgroup to improve with a decrease in coverage. We prove that if a feature representation satisfies the sufficiency criterion or is calibrated for mean and variance, than the proposed fairness criteria is met. Further, we introduce two approaches to mitigate the performance disparity across subgroups: (a) by regularizing an upper bound of conditional mutual information under a Gaussian assumption and (b) by regularizing a contrastive loss for conditional mean and conditional variance prediction. The effectiveness of these approaches is demonstrated on synthetic and real-world datasets.
翻译:如果对准确预测的信心不足,则选择性回归可以避免预测。 一般来说,通过允许拒绝选项,人们预计回归模型的性能会以降低覆盖率(即预测较少的样本)为代价而增加。 但是,正如我们所显示的,在某些情况下,少数分组的性能会减少,而我们则会减少覆盖面,因此选择性回归会扩大不同敏感分组之间的差别。受这些差异的驱使,我们提出了选择性回归的新公平标准,要求每个分组的性能随着覆盖面的减少而提高。我们证明,如果特征代表符合充足标准,或者根据平均和差异加以校准,则不符合拟议的公平标准。此外,我们引入了两种办法来缩小各分组的性能差异:(a) 在高斯假设下将有条件的相互信息的上限正规化,以及(b) 将有条件的中值和有条件的差异预测定期化,这些方法的有效性在合成数据和现实世界数据集中得到证明。