Despite the success of deep-learning models in many tasks, there have been concerns about such models learning shortcuts, and their lack of robustness to irrelevant confounders. When it comes to models directly trained on human faces, a sensitive confounder is that of human identities. Many face-related tasks should ideally be identity-independent, and perform uniformly across different individuals (i.e. be fair). One way to measure and enforce such robustness and performance uniformity is through enforcing it during training, assuming identity-related information is available at scale. However, due to privacy concerns and also the cost of collecting such information, this is often not the case, and most face datasets simply contain input images and their corresponding task-related labels. Thus, improving identity-related robustness without the need for such annotations is of great importance. Here, we explore using face-recognition embedding vectors, as proxies for identities, to enforce such robustness. We propose to use the structure in the face-recognition embedding space, to implicitly emphasize rare samples within each class. We do so by weighting samples according to their conditional inverse density (CID) in the proxy embedding space. Our experiments suggest that such a simple sample weighting scheme, not only improves the training robustness, it often improves the overall performance as a result of such robustness. We also show that employing such constraints during training results in models that are significantly less sensitive to different levels of bias in the dataset.
翻译:尽管深度学习模型在许多任务中取得了成功,但也存在关于这些模型学习捷径和对无关混淆因素缺乏鲁棒性的担忧。当谈到直接训练人脸的模型时,敏感的混淆因素之一是人的身份。许多与人脸相关的任务理想情况下应该是身份无关的,并且在不同个体之间表现出一致性(即公平)。衡量和强制此类鲁棒性和性能一致性的一种方法是在训练过程中强制执行它,假设可以大规模地获得与身份相关的信息。然而,由于隐私问题以及收集此类信息的成本,通常情况下并不是这样,大多数人脸数据集仅包含输入图像及其相应的任务相关标签。因此,在不需要这些注释的情况下改进与身份相关的鲁棒性非常重要。这里,我们探讨使用人脸识别嵌入向量作为身份的替代品来强制执行这种鲁棒性。我们建议使用人脸识别嵌入空间中的结构,通过在每个类中根据其在代理嵌入空间中的条件倒数密度(CID)加权样本,隐式强调罕见样本。我们的实验表明,这样一个简单的加权样本方案不仅改善了训练的鲁棒性,而且通常由于这种鲁棒性的结果而改善了总体性能。我们还表明,在训练过程中采用这种约束条件会产生对数据集中不同程度偏差显著不敏感的模型。