Training semantic segmentation models requires a large amount of finely annotated data, making it hard to quickly adapt to novel classes not satisfying this condition. Few-Shot Segmentation (FS-Seg) tackles this problem with many constraints. In this paper, we introduce a new benchmark, called Generalized Few-Shot Semantic Segmentation (GFS-Seg), to analyze the generalization ability of segmentation models to simultaneously recognize novel categories with very few examples as well as base categories with sufficient examples. Previous state-of-the-art FS-Seg methods fall short in GFS-Seg and the performance discrepancy mainly comes from the constrained training setting of FS-Seg. To make GFS-Seg tractable, we set up a GFS-Seg baseline that achieves decent performance without structural change on the original model. Then, as context is the key for boosting performance on semantic segmentation, we propose the Context-Aware Prototype Learning (CAPL) that significantly improves performance by leveraging the contextual information to update class prototypes with aligned features. Extensive experiments on Pascal-VOC and COCO manifest the effectiveness of CAPL, and CAPL also generalizes well to FS-Seg.
翻译:培训语义分解模型需要大量精细附加说明的数据,因此很难快速适应不符合此条件的新类。 少点偏偏分割(FS-Seg)用许多限制来解决这个问题。 在本文件中,我们引入了一个新的基准,称为“通用的少点静默分解(GFS-Seg) ”,以分析分解模型的概括能力,从而同时识别新分类,并举几个例子和有足够实例的基类。 以往最先进的FS-Seg方法在GFS-Seg中落后于GFS-Seg, 性能差异主要来自FS-Seg的有限培训设置。为了使GFS-Seg可调整,我们设置了一个GFS-Seg基准,在不改变原始模型结构的情况下实现体面业绩。 然后,作为提高语义分解功能的关键,我们建议使用背景软件Prototylein(CAPL),通过利用背景信息更新教室原型模型来显著改进性能,同时更新Pascal-VOC和COS-CAPL的通用试验。