In recent works, utilizing a deep network trained on meta-training set serves as a strong baseline in few-shot learning. In this paper, we move forward to refine novel-class features by finetuning a trained deep network. Finetuning is designed to focus on reducing biases in novel-class feature distributions, which we define as two aspects: class-agnostic and class-specific biases. Class-agnostic bias is defined as the distribution shifting introduced by domain difference, which we propose Distribution Calibration Module(DCM) to reduce. DCM owes good property of eliminating domain difference and fast feature adaptation during optimization. Class-specific bias is defined as the biased estimation using a few samples in novel classes, which we propose Selected Sampling(SS) to reduce. Without inferring the actual class distribution, SS is designed by running sampling using proposal distributions around support-set samples. By powering finetuning with DCM and SS, we achieve state-of-the-art results on Meta-Dataset with consistent performance boosts over ten datasets from different domains. We believe our simple yet effective method demonstrates its possibility to be applied on practical few-shot applications.
翻译:在最近的著作中,利用经过元培训训练的深网络,在几张短片的学习中,成为了强有力的基准。在本文中,我们通过对经过训练的深层网络进行微调,着手完善小类特征特征。微调的目的是侧重于减少小类特征分布中的偏见,我们将其定义为两个方面:阶级认知和阶级特有偏见。分类认知偏向被定义为按域差异引入的分布变化,我们提议缩小分布校准模块(DCM ) 。 DCM 拥有在优化期间消除域差异和快速特征适应的良好属性。 类别偏向被定义为利用新类中的少数样本进行偏向性估计,我们建议减少这些样本。 在不推断实际的类别分布的情况下,SS 是通过使用支持集样本的分布提案进行抽样设计。 通过对 DCM 和 SS 进行微调,我们实现了Meta-Dataset的状态艺术结果, 并同时对来自不同领域的十个数据集进行连续的性能增强。我们认为,我们简单而有效的方法表明它有可能适用于实际的少数应用。