Few-shot learning addresses the challenge of learning how to address novel tasks given not just limited supervision but limited data as well. An attractive solution is synthetic data generation. However, most such methods are overly sophisticated, focusing on high-quality, realistic data in the input space. It is unclear whether adapting them to the few-shot regime and using them for the downstream task of classification is the right approach. Previous works on synthetic data generation for few-shot classification focus on exploiting complex models, e.g. a Wasserstein GAN with multiple regularizers or a network that transfers latent diversities from known to novel classes. We follow a different approach and investigate how a simple and straightforward synthetic data generation method can be used effectively. We make two contributions, namely we show that: (1) using a simple loss function is more than enough for training a feature generator in the few-shot setting; and (2) learning to generate tensor features instead of vector features is superior. Extensive experiments on miniImagenet, CUB and CIFAR-FS datasets show that our method sets a new state of the art, outperforming more sophisticated few-shot data augmentation methods. The source code can be found at https://github.com/MichalisLazarou/TFH_fewshot.
翻译:少见的学习解决了学习如何解决新任务的挑战,不仅因为监督有限,而且数据也有限。一个有吸引力的解决办法是合成数据生成。然而,大多数这类方法过于复杂,侧重于输入空间的高质量、现实的数据,不清楚是否正确。以前关于合成数据生成的合成数据生成工作侧重于利用复杂模型,例如,拥有多个正规化器的瓦西尔斯泰因GAN或将已知的潜在多样性转移至新类的网络。我们采取不同的做法,调查如何有效使用简单、直接的合成数据生成方法。我们作出两种贡献,即:(1) 使用简单的损失功能足以在几发环境中培训一个特性生成器;(2) 学习生成抗声特性而不是矢量特性是优越的。关于小型IMagenet、CUB和CIFAR-FS数据集的广泛实验表明,我们的方法设置了新的艺术状态,比更精密的少量数据扩展方法要差得多。我们在http://Mlisqualus/Mishalushoma/Mishalmasroy 上找到源码。