Small data challenges have emerged in many learning problems, since the success of deep neural networks often relies on the availability of a huge amount of labeled data that is expensive to collect. To address it, many efforts have been made on training complex models with small data in an unsupervised and semi-supervised fashion. In this paper, we will review the recent progresses on these two major categories of methods. A wide spectrum of small data models will be categorized in a big picture, where we will show how they interplay with each other to motivate explorations of new ideas. We will review the criteria of learning the transformation equivariant, disentangled, self-supervised and semi-supervised representations, which underpin the foundations of recent developments. Many instantiations of unsupervised and semi-supervised generative models have been developed on the basis of these criteria, greatly expanding the territory of existing autoencoders, generative adversarial nets (GANs) and other deep networks by exploring the distribution of unlabeled data for more powerful representations. While we focus on the unsupervised and semi-supervised methods, we will also provide a broader review of other emerging topics, from unsupervised and semi-supervised domain adaptation to the fundamental roles of transformation equivariance and invariance in training a wide spectrum of deep networks. It is impossible for us to write an exclusive encyclopedia to include all related works. Instead, we aim at exploring the main ideas, principles and methods in this area to reveal where we are heading on the journey towards addressing the small data challenges in this big data era.
翻译:在许多学习问题中出现了小型数据挑战,因为深层神经网络的成功往往取决于大量标签数据能否成功,而这些数据收集费用昂贵。为了解决这个问题,已经作出许多努力,以不受监督和半监督的方式培训具有小数据的复杂模型。在本文件中,我们将审查这两大类方法的最新进展。一系列小型数据模型将在大图中进行分类,我们将展示它们如何相互作用,以激励探索新的思想。我们将审查学习变异、混乱、自我监督、半监督的表达的标准,这是最近发展的基础。许多未经监督和半监督的变异模型,根据这些标准,我们将审查这两大类方法的最新进展。将大大扩大现有自动化网络、基因化的对抗网(GANs)和其他深层网络的范围,探索如何将未标记的数据分布用于更强大的表达。我们将研究在这种变异异、非反复的变异变、非反复、自我监督和半监督的表达方法中,我们将从这个变异变的变的轨道上,再审视这个变异的模型,然后再审视其他的变异性模型,然后再审视其他的变异性模型。