Visual sentiment analysis has received increasing attention in recent years. However, the quality of the dataset is a concern because the sentiment labels are crowd-sourcing, subjective, and prone to mistakes. This poses a severe threat to the data-driven models including the deep neural networks which would generalize poorly on the testing cases if they are trained to over-fit the samples with noisy sentiment labels. Inspired by the recent progress on learning with noisy labels, we propose a robust learning method to perform robust visual sentiment analysis. Our method relies on an external memory to aggregate and filter noisy labels during training and thus can prevent the model from overfitting the noisy cases. The memory is composed of the prototypes with corresponding labels, both of which can be updated online. We establish a benchmark for visual sentiment analysis with label noise using publicly available datasets. The experiment results of the proposed benchmark settings comprehensively show the effectiveness of our method.
翻译:视觉感知分析近年来受到越来越多的关注。然而,数据集的质量是一个令人关切的问题,因为情绪标签是众包、主观和容易出错的。这对数据驱动模型构成了严重威胁,包括深神经网络,如果它们受过训练,能够用吵闹情绪标签为样本超标,测试案例就会普遍不甚普遍。由于最近使用吵闹标签学习的进展,我们提出了一个强有力的学习方法,以进行强健的视觉感知分析。我们的方法依靠外部记忆来汇总和过滤吵闹标签,从而可以防止模型过度适应吵闹案例。记忆由带有相应标签的原型组成,两者都可以在网上更新。我们用公开的数据集建立一个视觉感知分析基准,用标签噪音来建立基准。拟议基准设置的实验结果全面显示了我们方法的有效性。