Existing few-shot learning (FSL) methods rely on training with a large labeled dataset, which prevents them from leveraging abundant unlabeled data. From an information-theoretic perspective, we propose an effective unsupervised FSL method, learning representations with self-supervision. Following the InfoMax principle, our method learns comprehensive representations by capturing the intrinsic structure of the data. Specifically, we maximize the mutual information (MI) of instances and their representations with a low-bias MI estimator to perform self-supervised pre-training. Rather than supervised pre-training focusing on the discriminable features of the seen classes, our self-supervised model has less bias toward the seen classes, resulting in better generalization for unseen classes. We explain that supervised pre-training and self-supervised pre-training are actually maximizing different MI objectives. Extensive experiments are further conducted to analyze their FSL performance with various training settings. Surprisingly, the results show that self-supervised pre-training can outperform supervised pre-training under the appropriate conditions. Compared with state-of-the-art FSL methods, our approach achieves comparable performance on widely used FSL benchmarks without any labels of the base classes.
翻译:现有的微小学习(FSL)方法依赖于使用大标签数据集的培训,这使它们无法利用大量未贴标签的数据。从信息理论的角度来看,我们建议一种有效的、不受监督的FSL方法,学习自我监督的演示。根据InfoMax原则,我们的方法通过捕捉数据的内在结构学习了全面的表述方式。具体地说,我们最大限度地利用实例的相互信息及其以低偏差的MI测量器进行自我监督的预培训。我们自我监督的模型不是以所见的班级的不同特征为主的监督前培训,而是对所见班级的偏向性较低,从而导致对隐蔽班级的更普遍化。我们解释说,受监督的预培训和自我监督的预培训实际上正在最大限度地实现不同的MI目标。我们进一步进行了广泛的实验,以利用各种培训环境来分析其FSL的绩效。令人惊讶的是,结果显示,自我监督的预培训可以在适当条件下超越监督的预培训。与FSFSL方法的州级基准相比,没有使用的任何可比的FSFSL方法。