Deep models are susceptible to learning spurious correlations, even during the post-processing. We take a closer look at the knowledge distillation -- a popular post-processing technique for model compression -- and find that distilling with biased training data gives rise to a biased student, even when the teacher is debiased. To address this issue, we propose a simple knowledge distillation algorithm, coined DeTT (Debiasing by Teacher Transplanting). Inspired by a recent observation that the last neural net layer plays an overwhelmingly important role in debiasing, DeTT directly transplants the teacher's last layer to the student. Remaining layers are distilled by matching the feature map outputs of the student and the teacher, where the samples are reweighted to mitigate the dataset bias. Importantly, DeTT does not rely on the availability of extensive annotations on the bias-related attribute, which is typically not available during the post-processing phase. Throughout our experiments, DeTT successfully debiases the student model, consistently outperforming the baselines in terms of the worst-group accuracy.
翻译:深层模型很容易学习假的关联关系,即使在后处理期间也是如此。我们更仔细地研究知识蒸馏过程 -- -- 一种流行的模型压缩后处理技术 -- -- 并发现,用偏颇的培训数据蒸馏产生偏向性学生,即使教师被贬低。为了解决这个问题,我们提议了一个简单的知识蒸馏算法,即教师移植的偏见催化法。最近观察到最后一个神经网层在贬低偏向方面起着极其重要的作用,我们发现,最后的神经网层将教师的最后一层直接移植给学生。其余层通过匹配学生和教师的地貌图输出来蒸馏。通过对样本进行重新加权来减轻数据设置偏向性来蒸馏。重要的是,DTT并不依赖与偏向有关属性的广泛说明,而后者通常在后处理阶段是无法获得的。在我们进行的实验中,DTT成功地降低了学生模型的偏向性,在最差的精确度方面一直超过基线。