Virtually all of deep learning literature relies on the assumption of large amounts of available training data. Indeed, even the majority of few-shot learning methods rely on a large set of "base classes" for pretraining. This assumption, however, does not always hold. For some tasks, annotating a large number of classes can be infeasible, and even collecting the images themselves can be a challenge in some scenarios. In this paper, we study this problem and call it "Small Data" setting, in contrast to "Big Data". To unlock the full potential of small data, we propose to augment the models with annotations for other related tasks, thus increasing their generalization abilities. In particular, we use the richly annotated scene parsing dataset ADE20K to construct our realistic Long-tail Recognition with Diverse Supervision (LRDS) benchmark by splitting the object categories into head and tail based on their distribution. Following the standard few-shot learning protocol, we use the head classes for representation learning and the tail classes for evaluation. Moreover, we further subsample the head categories and images to generate two novel settings which we call "Scarce-Class" and "Scarce-Image", respectively corresponding to the shortage of samples for rare classes and training images. Finally, we analyze the effect of applying various additional supervision sources under the proposed settings. Our experiments demonstrate that densely labeling a small set of images can indeed largely remedy the small data constraints.
翻译:事实上,几乎所有深层次的学习文献都依赖于大量现有培训数据的假设。事实上,即使大多数少见的学习方法都依赖于大量的“基础班”来进行预培训。但是,这一假设并不总能维持。对于某些任务,大量分类的说明可能不可行,甚至收集图像本身也在某些情景中是一个挑战。在本文中,我们研究这一问题,并称之为“小数据”设置。为了释放小数据的全部潜力,我们提议用其他相关任务的说明来充实模型,从而增强它们的概括化能力。特别是,我们使用大量附加注释的场景对数据集ADE20K进行解析,以构建现实的长尾类识别,根据对象的分布将对象类别分成头部和尾部。我们根据标准的“少见学习协议”,我们使用拟议的头类来进行演示学习,用尾类来评估。此外,我们进一步对头类和图像进行补充,以产生两个新的环境,我们称之为“Scar20K ”, 高端图像的解析,我们最后用“小的样本” 和“小类” 来展示“我们“小的标签” 的“小分析“ ” 的标签” 。我们最后的“小的标签”,可以显示“小的“小的标签” 的标签” 。