In this paper, we introduce a new architecture for few shot learning, the task of teaching a neural network from as few as one or five labeled examples. Inspired by the theoretical results of Alaine et al that Denoising Autoencoders refine features to lie closer to the true data manifold, we present a new training scheme that adds noise at multiple stages of an existing neural architecture while simultaneously learning to be robust to this added noise. This architecture, which we call a Self-Denoising Neural Network (SDNN), can be applied easily to most modern convolutional neural architectures, and can be used as a supplement to many existing few-shot learning techniques. We empirically show that SDNNs out-perform previous state-of-the-art methods for few shot image recognition using the Wide-ResNet architecture on the \textit{mini}ImageNet, tiered-ImageNet, and CIFAR-FS few shot learning datasets. We also perform a series of ablation experiments to empirically justify the construction of the SDNN architecture. Finally, we show that SDNNs even improve few shot performance on the task of human action detection in video using experiments on the ActEV SDL Surprise Activities challenge.
翻译:在本文中,我们引入了用于少点拍摄学习的新架构,即从少到少到少教授神经神经网络的任务。在Alane 等人的理论结果的启发下,Denoising Autoenccoders 改进了一些功能,使之更接近真正的数据元,我们展示了一个新的培训计划,在现有的神经结构的多个阶段增加噪音,同时学习对添加的噪音的强大。这个我们称之为自我 Denoizing神经网络(SDNNN)的架构可以很容易地应用于大多数现代神经神经神经结构(SDNN),并且可以用来补充许多现有的少见的学习技术。我们从经验上表明,SDNNNS超越了以前最先进的微小图像识别方法,在\textit{mini}IMageNet、分级-IMageNet和CIFAR-FS少片段的学习数据集中,我们进行了一系列实验,以便从经验上证明建造SDNNN结构是合理的。最后,我们展示SDNDNNS甚至改进了在人类探测任务中进行SD VIV行动的挑战。