This paper introduces a generalized few-shot segmentation framework with a straightforward training process and an easy-to-optimize inference phase. In particular, we propose a simple yet effective model based on the well-known InfoMax principle, where the Mutual Information (MI) between the learned feature representations and their corresponding predictions is maximized. In addition, the terms derived from our MI-based formulation are coupled with a knowledge distillation term to retain the knowledge on base classes. With a simple training process, our inference model can be applied on top of any segmentation network trained on base classes. The proposed inference yields substantial improvements on the popular few-shot segmentation benchmarks PASCAL-$5^i$ and COCO-$20^i$. Particularly, for novel classes, the improvement gains range from 5% to 20% (PASCAL-$5^i$) and from 2.5% to 10.5% (COCO-$20^i$) in the 1-shot and 5-shot scenarios, respectively. Furthermore, we propose a more challenging setting, where performance gaps are further exacerbated. Our code is publicly available at https://github.com/sinahmr/DIaM.
翻译:本文提出了一个通用的微小分解框架,其培训过程直截了当,并且容易优化推导阶段,特别是,我们根据众所周知的信息-最大原则,提出一个简单而有效的模式,即所学地貌说明及其相应预测之间的相互信息最大化;此外,我们以MI为基础的提法中所产生的术语加上一个知识提炼术语,以保留基础班的知识;通过一个简单的培训过程,我们的推论模型可以适用于在基础班上培训的任何分解网络之上;提议的推论在流行的微分解基准PASCAL-5美元和CO-20美元上取得了实质性的改进。特别是,在新类中,改进的收益从5%到20%(PASCAL-5美元)不等,在1集和5集的情景中分别从2.5%到10.5%(CO-20美元)不等。此外,我们提议了更具有挑战性的设置,因为业绩差距会进一步加剧。我们的代码可在 https://ghmur DI上公开查阅。