Encouraging progress in few-shot semantic segmentation has been made by leveraging features learned upon base classes with sufficient training data to represent novel classes with few-shot examples. However, this feature sharing mechanism inevitably causes semantic aliasing between novel classes when they have similar compositions of semantic concepts. In this paper, we reformulate few-shot segmentation as a semantic reconstruction problem, and convert base class features into a series of basis vectors which span a class-level semantic space for novel class reconstruction. By introducing contrastive loss, we maximize the orthogonality of basis vectors while minimizing semantic aliasing between classes. Within the reconstructed representation space, we further suppress interference from other classes by projecting query features to the support vector for precise semantic activation. Our proposed approach, referred to as anti-aliasing semantic reconstruction (ASR), provides a systematic yet interpretable solution for few-shot learning problems. Extensive experiments on PASCAL VOC and MS COCO datasets show that ASR achieves strong results compared with the prior works.
翻译:通过利用在基础班上学习的具有足够培训数据、代表新班的培训数据并举几个例子,在几发语义分解方面取得了令人鼓舞的进展。然而,这种特征分享机制不可避免地导致新班在具有类似语义概念构成的新班之间的语义别名。在本文中,我们重新将几发语义分解作为一个语义重建问题,并将基级特性转换成一系列基础矢量,这些矢量跨越了等级层次的语义空间,用于新班级重建。通过引入对比性损失,我们最大限度地扩大基级矢量的正交错性,同时尽量减少班级之间的语义化别名。在重建的代表空间内,我们进一步抑制来自其他班级的语义别名,方法是为精确的语义激活向支持矢量投放查询特征。我们提出的方法,即反言语义重建(ASR),为几发语义学习问题提供了一个系统但又可解释的解决办法。关于PASAL VOC 和MS CO 数据集的广泛实验显示,与先前的工程相比,ASR取得了强烈的结果。