Few-shot and one-shot learning have been the subject of active and intensive research in recent years, with mounting evidence pointing to successful implementation and exploitation of few-shot learning algorithms in practice. Classical statistical learning theories do not fully explain why few- or one-shot learning is at all possible since traditional generalisation bounds normally require large training and testing samples to be meaningful. This sharply contrasts with numerous examples of successful one- and few-shot learning systems and applications. In this work we present mathematical foundations for a theory of one-shot and few-shot learning and reveal conditions specifying when such learning schemes are likely to succeed. Our theory is based on intrinsic properties of high-dimensional spaces. We show that if the ambient or latent decision space of a learning machine is sufficiently high-dimensional than a large class of objects in this space can indeed be easily learned from few examples provided that certain data non-concentration conditions are met.
翻译:近些年来,对少发和一发学习进行了积极和深入的研究,越来越多的证据表明在实践中成功实施和利用了少发学习算法,古典统计学理论没有完全解释为什么可以进行少发或一发学习,因为传统的概括性界限通常要求大量的培训和测试样本才有意义,这与一发和少发学习系统和应用的成功事例大相径庭。在这项工作中,我们为一发和少发学习理论提供了数学基础,并揭示了可能取得成功的条件。我们的理论以高维空间的内在特性为基础。我们表明,如果学习机器的环境或潜在决策空间比这一空间的一大批物体具有足够高的维度,确实可以很容易地从少数例子中学习,只要满足某些非集中性数据条件即可。