Semi-supervised learning (SSL) is an efficient framework that can train models with both labeled and unlabeled data, but may generate ambiguous and non-distinguishable representations when lacking adequate labeled samples. With human-in-the-loop, active learning can iteratively select informative unlabeled samples for labeling and training to improve the performance in the SSL framework. However, most existing active learning approaches rely on pre-trained features, which is not suitable for end-to-end learning. To deal with the drawbacks of SSL, in this paper, we propose a novel end-to-end representation learning method, namely ActiveMatch, which combines SSL with contrastive learning and active learning to fully leverage the limited labels. Starting from a small amount of labeled data with unsupervised contrastive learning as a warm-up, ActiveMatch then combines SSL and supervised contrastive learning, and actively selects the most representative samples for labeling during the training, resulting in better representations towards the classification. Compared with MixMatch and FixMatch with the same amount of labeled data, we show that ActiveMatch achieves the state-of-the-art performance, with 89.24% accuracy on CIFAR-10 with 100 collected labels, and 92.20% accuracy with 200 collected labels.
翻译:半监督的学习(SSL)是一个有效的框架,它可以用标签和未标签的数据来培训模型,但是在缺乏适当的标签样本时,可能会产生模糊和不可区分的表达方式。在人行中,积极学习可以迭接地选择信息性、无标签的样本,用于标签和培训,以改进SSL框架的性能。然而,大多数现有的积极学习方法都依靠预先培训的特征,不适合端到端学习。为了处理SSL的缺陷,我们在本文件中提出了一个新的端到端代表学习方法,即“主动匹配”,将SSL与对比学习和积极学习相结合,以充分利用有限的标签。从少量带有非超强对比学习的标签样本开始,作为温暖学习,主动Match然后将SSL与监督的对比学习结合起来,积极选择培训期间最有代表性的标签样本,从而更好地进行分类。将MixMatch和固定匹配与相同数量的标签的准确度结合起来,92。我们从少量标签采集的标签数据开始,与100-10的准确度,我们显示“积极匹配”与100的标签-10的准确性业绩,我们显示在收集的“积极匹配”