Few-shot segmentation has recently attracted substantial interest, with the popular meta-learning paradigm widely dominating the literature. We show that the way inference is performed for a given few-shot segmentation task has a substantial effect on performances, an aspect that has been overlooked in the literature. We introduce a transductive inference, which leverages the statistics of the unlabeled pixels of a task by optimizing a new loss containing three complementary terms: (i) a standard cross-entropy on the labeled pixels; (ii) the entropy of posteriors on the unlabeled query pixels; and (iii) a global KL-divergence regularizer based on the proportion of the predicted foreground region. Our inference uses a simple linear classifier of the extracted features, has a computational load comparable to inductive inference and can be used on top of any base training. Using standard cross-entropy training on the base classes, our inference yields highly competitive performances on well-known few-shot segmentation benchmarks. On PASCAL-5i, it brings about 5% improvement over the best performing state-of-the-art method in the 5-shot scenario, while being on par in the 1-shot setting. Even more surprisingly, this gap widens as the number of support samples increases, reaching up to 6% in the 10-shot scenario. Furthermore, we introduce a more realistic setting with domain shift, where the base and novel classes are drawn from different datasets. In this setting, we found that our method achieves the best performances.
翻译:微小的分解最近引起了极大的兴趣, 流行的元化学习模式在文献中占据了主导地位。 我们显示, 如何对特定微小分解任务进行推论, 对业绩产生了重大影响, 文献中忽视了这一方面。 我们引入了转导推论, 通过优化含有三个补充术语的新损失, 利用未贴标签的像素的未贴标签像素等像素的统计, 使任务中一个未贴标签像素的像素的像素的统计值产生杠杆效应。 (一) 标签像素上的标准交叉吸附; (二) 未贴标签查询像素上的后部; (三) 基于预测的地平面区域比例, 全球KL- 振动常量常量常量常量常量对业绩有影响。 我们的推导出一个简单的直线分解器, 与任何基本训练的精度相仿, 。 使用标准交叉吸附训练, 我们的分解法在已知的微小分解基准上产生高度的竞争性的表现。 在PASACL-5 中, 更精确的推算中, 我们的推算出了一个最接近的推算, 推算了最接近的推算了最接近的推算法 。 。