In this paper, we propose a Generative Translation Classification Network (GTCN) for improving visual classification accuracy in settings where classes are visually similar and data is scarce. For this purpose, we propose joint learning from a scratch to train a classifier and a generative stochastic translation network end-to-end. The translation network is used to perform on-line data augmentation across classes, whereas previous works have mostly involved domain adaptation. To help the model further benefit from this data-augmentation, we introduce an adaptive fade-in loss and a quadruplet loss. We perform experiments on multiple datasets to demonstrate the proposed method's performance in varied settings. Of particular interest, training on 40% of the dataset is enough for our model to surpass the performance of baselines trained on the full dataset. When our architecture is trained on the full dataset, we achieve comparable performance with state-of-the-art methods despite using a light-weight architecture.
翻译:在本文中,我们建议建立一个“创用翻译分类网络”,以提高各类视觉相似且数据稀少的环境的视觉分类准确性。 为此,我们建议从零开始联合学习,以训练一个分类者和基因切换翻译网络的端到端。翻译网络用于在各类进行在线数据扩增,而以前的工程大多涉及域适应。为了帮助模型进一步受益于这一数据增强,我们引入了适应性淡化损失和四分位损失。我们在多个数据集上进行了实验,以展示拟议方法在不同环境中的性能。特别令人感兴趣的是,40%的数据集培训足以使我们的模型超过在全数据集方面培训的基线性能。当我们的架构接受全数据集培训时,我们尽管使用了轻量的架构,但我们还是取得了与最新技术方法的可比性能。