Currently, the state-of-the-art methods treat few-shot semantic segmentation task as a conditional foreground-background segmentation problem, assuming each class is independent. In this paper, we introduce the concept of meta-class, which is the meta information (e.g. certain middle-level features) shareable among all classes. To explicitly learn meta-class representations in few-shot segmentation task, we propose a novel Meta-class Memory based few-shot segmentation method (MM-Net), where we introduce a set of learnable memory embeddings to memorize the meta-class information during the base class training and transfer to novel classes during the inference stage. Moreover, for the $k$-shot scenario, we propose a novel image quality measurement module to select images from the set of support images. A high-quality class prototype could be obtained with the weighted sum of support image features based on the quality measure. Experiments on both PASCAL-$5^i$ and COCO dataset shows that our proposed method is able to achieve state-of-the-art results in both 1-shot and 5-shot settings. Particularly, our proposed MM-Net achieves 37.5\% mIoU on the COCO dataset in 1-shot setting, which is 5.1\% higher than the previous state-of-the-art.
翻译:目前,最先进的方法将微粒语义分解任务作为有条件的地表-地表分解问题处理,假设每类都是独立的。在本文中,我们引入了元类概念,即所有类都可以共享的元信息(例如某些中等级特征);为了明确学习以微粒分解任务为基础的元类表示,我们建议采用基于微粒分解任务的新颖的元类记忆分解方法(MM-Net),在基础班培训期间引入一套可学习的内存嵌入,将元类信息混为一流,并在推断阶段将信息转移到新类。此外,对于美元分解情景,我们建议采用一个新的图像质量计量模块,从成套支持图像中选择图像(例如某些中等级特征),用基于微粒分解任务的支持图像特征的加权组合获得高质量的类原型。对PASAL-5美元和COCO数据集的实验表明,我们拟议的方法能够在1张地图和5MO-M-O-M-O-M-M-MS-MS-G-G-MS-MS-MS-G-MS-MS-MS-MS-MS-MS-MS-G-MS-MS-MS-MS-M-MS-MS-MS-MS-N-MS-M-M-M-MS-MS-M-M-M-M-M-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G-G