Few-shot classification addresses the challenge of classifying examples given not just limited supervision but limited data as well. An attractive solution is synthetic data generation. However, most such methods are overly sophisticated, focusing on high-quality, realistic data in the input space. It is unclear whether adapting them to the few-shot regime and using them for the downstream task of classification is the right approach. Previous works on synthetic data generation for few-shot classification focus on exploiting complex models, e.g. a Wasserstein GAN with multiple regularizers or a network that transfers latent diversities from known to novel classes. We follow a different approach and investigate how a simple and straightforward synthetic data generation method can be used effectively. We make two contributions, namely we show that: (1) using a simple loss function is more than enough for training a feature generator in the few-shot setting; and (2) learning to generate tensor features instead of vector features is superior. Extensive experiments on miniImagenet, CUB and CIFAR-FS datasets show that our method sets a new state of the art, outperforming more sophisticated few-shot data augmentation methods.
翻译:微小的分类处理不仅有限监督,而且数据也有限对实例进行分类的挑战。一个有吸引力的解决办法是合成数据生成。然而,大多数这类方法过于复杂,侧重于输入空间的高质量、现实的数据。不清楚的是,正确的方法是否使这些方法适应微小的系统,并将其用于下游分类任务。以前为微小的分类进行的合成数据生成工作侧重于利用复杂模型,例如,拥有多个正规化器的瓦西尔斯坦GAN或将已知的潜在多样性转移至新类的网络。我们采取不同的做法,并调查如何有效使用简单、直接的合成数据生成方法。我们做出两种贡献,即:(1) 使用简单的损失功能足以在微小的环境下培训特性生成器;(2) 学习生成高压特性而不是矢量特性是优越的。关于微型Imagenet、CUB和CIFAR-FS数据集的广泛实验表明,我们的方法提出了新的艺术状态,比更复杂的微小的增强数据生成方法要差。