In supervised learning - for instance in image classification - modern massive datasets are commonly labeled by a crowd of workers. The obtained labels in this crowdsourcing setting are then aggregated for training. The aggregation step generally leverages a per-worker trust score. Yet, such worker-centric approaches discard each task's ambiguity. Some intrinsically ambiguous tasks might even fool expert workers, which could eventually be harmful to the learning step. In a standard supervised learning setting - with one label per task - the Area Under the Margin (AUM) is tailored to identify mislabeled data. We adapt the AUM to identify ambiguous tasks in crowdsourced learning scenarios, introducing the Weighted AUM (WAUM). The WAUM is an average of AUMs weighted by task-dependent scores. We show that the WAUM can help discard ambiguous tasks from the training set, leading to better generalization or calibration performance. We report improvements over existing strategies for learning a crowd, both for simulated settings and for the CIFAR-10H, LabelMe and Music crowdsourced datasets.
翻译:在监督的学习中(例如在图像分类中),现代大规模数据集通常由一群工人贴上标签,然后将这种众包环境中获得的标签汇总起来,以供培训使用。聚合步骤通常会利用每个工人的信用分数。然而,这种以工人为中心的方法抛弃了每一项任务的模糊性。一些内在的模棱两可的任务甚至会愚弄专家工人,最终会损害学习步骤。在一个标准的监督学习环境中(每个任务贴上一个标签),“马林之下地区(AUM)”是专门为识别标记错误的数据而设计的。我们调整了AUM,以辨别在多方来源的学习情景中含糊不清的任务,引入了WAUM(WAUM) 。WAUM是按任务加权的分数加权的AUMs平均数。我们表明,WAUM可以帮助从培训组中抛弃模糊性的任务,从而导致更好的概括性或校准性表现。我们报告了现有学习人群战略的改进情况,既用于模拟环境,也用于CIFAR-10H、LabelMe和音乐群源数据集。