Most existing few-shot learning (FSL) methods require a large amount of labeled data in meta-training, which is a major limit. To reduce the requirement of labels, a semi-supervised meta-training setting has been proposed for FSL, which includes only a few labeled samples and numbers of unlabeled samples in base classes. However, existing methods under this setting require class-aware sample selection from the unlabeled set, which violates the assumption of unlabeled set. In this paper, we propose a practical semi-supervised meta-training setting with truly unlabeled data. Under the new setting, the performance of existing methods drops notably. To better utilize both the labeled and truly unlabeled data, we propose a simple and effective meta-training framework, called pseudo-labeling based on meta-learning (PLML). Firstly, we train a classifier via common semi-supervised learning (SSL) and use it to obtain the pseudo-labels of unlabeled data. Then we build few-shot tasks from labeled and pseudo-labeled data and run meta-learning over the constructed tasks to learn the FSL model. Surprisingly, through extensive experiments across two FSL datasets, we find that this simple meta-training framework effectively prevents the performance degradation of FSL under limited labeled data. Besides, benefiting from meta-training, the proposed method improves the classifiers learned by two representative SSL algorithms as well.
翻译:多数现有的少许学习方法(FSL)要求大量在元培训中提供标签数据,这是一个重要的限制。为了减少标签要求,已经为FSL提议了一个半监督的元培训设置,其中仅包括少数标签样本和基类中未贴标签样本的数量。然而,在这种设置下的现有方法要求从未贴标签的数据集中选择类意识样本,这违反了未贴标签数据集的假设。在本文中,我们建议用真正未贴标签的数据建立一个实用的半监督的元培训设置。在新的设置下,现有方法的性能明显下降。为了更好地利用标签和真正未贴标签的数据,我们提议了一个简单有效的元培训框架,称为基于元学习(PLML)的假标签。首先,我们通过共同的半监督学习集成样本(SSL)来培训一个分类,用它来获取未贴标签数据的假标签。然后,我们从标签和假标签数据中建立少发的任务,然后在新设置的模板中进行代行的元学习。