Large-scale datasets are important for the development of deep learning models. Such datasets usually require a heavy workload of annotations, which are extremely time-consuming and expensive. To accelerate the annotation procedure, multiple annotators may be employed to label different subsets of the data. However, the inconsistency and bias among different annotators are harmful to the model training, especially for qualitative and subjective tasks.To address this challenge, in this paper, we propose a novel contrastive regression framework to address the disjoint annotations problem, where each sample is labeled by only one annotator and multiple annotators work on disjoint subsets of the data. To take account of both the intra-annotator consistency and inter-annotator inconsistency, two strategies are employed.Firstly, a contrastive-based loss is applied to learn the relative ranking among different samples of the same annotator, with the assumption that the ranking of samples from the same annotator is unanimous. Secondly, we apply the gradient reversal layer to learn robust representations that are invariant to different annotators. Experiments on the facial expression prediction task, as well as the image quality assessment task, verify the effectiveness of our proposed framework.
翻译:大型数据集对于发展深层学习模型十分重要。这类数据集通常需要繁重的注释工作量,这种工作量极为耗时和昂贵。为了加快批注程序,可以使用多批注员给数据的不同子集贴标签。然而,不同批注员之间的不一致和偏差有害于模式培训,特别是质量和主观任务。为了应对这一挑战,我们在本文件中提出了一个新的对比回归框架,以解决脱节的注释问题,因为每个样本只有一位注解员和多位注解员在数据分离子集上贴上标签。为了既考虑到批注员内部的一致性,又考虑到批注员之间的不一致,采用了两种策略。首先,采用基于对比的损失来学习同一批注员不同样本之间的相对排名,同时假设同一批注员的样本的排名是一致的。第二,我们使用梯度逆转层来学习不同批注员的稳健的描述。在面部表达预测任务上进行实验,以及作为图像质量评估的框架,核查我们提出的有效性。