Classifiers are biased when trained on biased datasets. As a remedy, we propose Learning to Split (ls), an algorithm for automatic bias detection. Given a dataset with input-label pairs, ls learns to split this dataset so that predictors trained on the training split cannot generalize to the testing split. This performance gap suggests that the testing split is under-represented in the dataset, which is a signal of potential bias. Identifying non-generalizable splits is challenging since we have no annotations about the bias. In this work, we show that the prediction correctness of each example in the testing split can be used as a source of weak supervision: generalization performance will drop if we move examples that are predicted correctly away from the testing split, leaving only those that are mis-predicted. ls is task-agnostic and can be applied to any supervised learning problem, ranging from natural language understanding and image classification to molecular property prediction. Empirical results show that ls is able to generate astonishingly challenging splits that correlate with human-identified biases. Moreover, we demonstrate that combining robust learning algorithms (such as group DRO) with splits identified by ls enables automatic de-biasing. Compared to previous state-of-the-art, we substantially improve the worst-group performance (23.4% on average) when the source of biases is unknown during training and validation.
翻译:在有偏差的数据集上培训时,分类者有偏差。作为一种补救措施,我们建议学习分解(ls),这是自动偏差检测的算法。在有输入标签配对的数据集中,我们学会了分解这个数据集,这样接受过培训的预测者就无法泛化为测试分解。这个性能差距表明,在数据集中,测试分解的代表性不足,这是潜在偏差的信号。确定不可概括的分解具有挑战性,因为我们没有关于偏差的说明。在这项工作中,我们显示测试分解中每个示例的预测正确性能可以用作薄弱的监督源:如果将预测正确脱离测试分解的示例移走,则一般化性能会下降,只留下那些被错误预知的预测者。这是一个任务性差异,可以应用于任何受监督的学习问题,从自然语言理解和图像分类到分子属性预测。根据实情结果显示,在人类分辨偏差的偏差中,将稳健的学习算法(例如最差的DRO组)与前位性平比(我们确认的平偏差)与前次的平比(我们确认的平比性) 使前次的成绩得以进行自动分析。