Few-shot recognition aims to recognize novel categories under low-data regimes. Due to the scarcity of images, machines cannot obtain enough effective information, and the generalization ability of the model is extremely weak. By using auxiliary semantic modalities, recent metric-learning based few-shot learning methods have achieved promising performances. However, these methods only augment the representations of support classes, while query images have no semantic modalities information to enhance representations. Instead, we propose attribute-shaped learning (ASL), which can normalize visual representations to predict attributes for query images. And we further devise an attribute-visual attention module (AVAM), which utilizes attributes to generate more discriminative features. Our method enables visual representations to focus on important regions with attributes guidance. Experiments demonstrate that our method can achieve competitive results on CUB and SUN benchmarks. Our code is available at {https://github.com/chenhaoxing/ASL}.
翻译:由于图像稀缺,机器无法获得足够有效的信息,模型的概括能力极弱。通过使用辅助语义模式,最近基于几发微光的标准化学习方法取得了有希望的成绩。然而,这些方法只能增加支持课程的表述,而查询图像没有语义模式信息来增强表达。相反,我们提议了属性形状学习(ASL),它可以使视觉表现正常化,以预测查询图像的属性。我们进一步设计了一个属性-视觉关注模块(AVAM),它利用属性产生更多的歧视特征。我们的方法使得视觉表现能够侧重于具有属性指导的重要区域。实验表明,我们的方法可以在CUB和SUN基准上取得竞争性结果。我们的代码可在 https://github.com/chenhaoxing/ASL}查阅。