The Zero-Shot Learning (ZSL) task attempts to learn concepts without any labeled data. Unlike traditional classification/detection tasks, the evaluation environment is provided unseen classes never encountered during training. As such, it remains both challenging, and promising on a variety of fronts, including unsupervised concept learning, domain adaptation, and dataset drift detection. Recently, there have been a variety of approaches towards solving ZSL, including improved metric learning methods, transfer learning, combinations of semantic and image domains using, e.g. word vectors, and generative models to model the latent space of known classes to classify unseen classes. We find many approaches require intensive training augmentation with attributes or features that may be commonly unavailable (attribute-based learning) or susceptible to adversarial attacks (generative learning). We propose combining approaches from the related person re-identification task for ZSL, with key modifications to ensure sufficiently improved performance in the ZSL setting without the need for feature or training dataset augmentation. We are able to achieve state-of-the-art performance on the CUB200 and Cars196 datasets in the ZSL setting compared to recent works, with NMI (normalized mutual inference) of 63.27 and top-1 of 61.04 for CUB200, and NMI 66.03 with top-1 82.75% in Cars196. We also show state-of-the-art results in the Generalized Zero-Shot Learning (GZSL) setting, with Harmonic Mean R-1 of 66.14% on the CUB200 dataset.
翻译:61. 与传统的分类/检测任务不同的是,评估环境提供了在培训期间从未遇到过的隐蔽课程。因此,评估环境仍然具有挑战性,而且在许多战线上都有希望,包括不受监督的概念学习、域适应和数据集漂移探测。最近,在解决 ZSL 方面采取了各种办法,包括改进衡量学习方法、转移学习、将语义和图像域结合起来而不使用任何标签数据。与传统的分类/检测任务不同,我们发现许多办法都需要强化培训,其属性或特征可能普遍得不到(基于属性的学习)或容易遭到对抗性攻击(基因学习)。我们建议将相关人员为 ZSL 重新定位任务采用的方法结合起来,同时进行重大修改,以确保充分改进ZSL 环境的性能,而不需要特征或培训数据集增强。 我们有能力在CUB200和Cars196类课的潜在空间中实现最先进的性能表现。