Few-shot segmentation aims to train a segmentation model that can fast adapt to novel classes with few exemplars. The conventional training paradigm is to learn to make predictions on query images conditioned on the features from support images. Previous methods only utilized the semantic-level prototypes of support images as conditional information. These methods cannot utilize all pixel-wise support information for the query predictions, which is however critical for the segmentation task. In this paper, we focus on utilizing pixel-wise relationships between support and query images to facilitate the few-shot segmentation task. We design a novel Cycle-Consistent TRansformer (CyCTR) module to aggregate pixel-wise support features into query ones. CyCTR performs cross-attention between features from different images, i.e. support and query images. We observe that there may exist unexpected irrelevant pixel-level support features. Directly performing cross-attention may aggregate these features from support to query and bias the query features. Thus, we propose using a novel cycle-consistent attention mechanism to filter out possible harmful support features and encourage query features to attend to the most informative pixels from support images. Experiments on all few-shot segmentation benchmarks demonstrate that our proposed CyCTR leads to remarkable improvement compared to previous state-of-the-art methods. Specifically, on Pascal-$5^i$ and COCO-$20^i$ datasets, we achieve 66.6% and 45.6% mIoU for 5-shot segmentation, outperforming previous state-of-the-art methods by 4.6% and 7.1% respectively.
翻译:少截截截面, 目的是训练一个可快速适应小类的分解模型。 常规培训模式是学习以支持图像的特性为条件对查询图像进行预测。 以往的方法只使用支持图像的语义级原型作为有条件信息。 这些方法无法使用所有像素支持的查询预测信息, 但对于分解任务来说却至关重要。 在本文件中, 我们侧重于使用支持和查询图像之间的等离子关系, 以方便微截面任务 。 我们设计了一个新型循环连接$$% TRansexer( CyCTR) 模块, 以综合像素支持功能为基础进行预测。 CyCTR对不同图像的特性进行交叉注意, 即支持和查询图像。 我们观察到可能存在出乎意料的离谱性像素级支持特性。 直接进行交叉访问, 可以从支持到查询和偏移方向查询特性。 因此, 我们提议使用一个新型循环连接关注机制, 来过滤超导值$$_ TR5, 将我们提出的最有害的支持功能- 演示到以前的直径分析方法。