Algorithmic bias often arises as a result of differential subgroup validity, in which predictive relationships vary across groups. For example, in toxic language detection, comments targeting different demographic groups can vary markedly across groups. In such settings, trained models can be dominated by the relationships that best fit the majority group, leading to disparate performance. We propose framing toxicity detection as multi-task learning (MTL), allowing a model to specialize on the relationships that are relevant to each demographic group while also leveraging shared properties across groups. With toxicity detection, each task corresponds to identifying toxicity against a particular demographic group. However, traditional MTL requires labels for all tasks to be present for every data point. To address this, we propose Conditional MTL (CondMTL), wherein only training examples relevant to the given demographic group are considered by the loss function. This lets us learn group specific representations in each branch which are not cross contaminated by irrelevant labels. Results on synthetic and real data show that using CondMTL improves predictive recall over various baselines in general and for the minority demographic group in particular, while having similar overall accuracy.
翻译:例如,在毒性语言检测中,针对不同人口群体的评论可以各群体之间差别很大。在这种环境下,经过培训的模型可以由最适合多数群体的关系主导,导致不同的性能。我们建议将毒性检测作为多任务学习(MTL)来设置,允许一种模型专门研究与每个人口群体有关的关系,同时利用不同群体之间的共同特性。通过毒性检测,每项任务对应于确定对特定人口群体的毒性。然而,传统的MTL要求为每个数据点的所有任务贴上标签。为了解决这个问题,我们提议采用条件性MTL(CondMTL)(CondMTL)(CondMTL),其中只考虑与特定人口群体有关的培训实例,而损失功能则考虑到这些实例。这让我们了解每个分支中未受到相关标签交叉污染的具体表现。合成和真实数据的结果显示,使用CondMTL(C)改进了对一般基线的预测性记忆,特别是少数群体人口群体的预测性,同时具有类似的总体准确性。