Learning accurate classifiers for novel categories from very few examples, known as few-shot image classification, is a challenging task in statistical machine learning and computer vision. The performance in few-shot classification suffers from the bias in the estimation of classifier parameters; however, an effective underlying bias reduction technique that could alleviate this issue in training few-shot classifiers has been overlooked. In this work, we demonstrate the effectiveness of Firth bias reduction in few-shot classification. Theoretically, Firth bias reduction removes the $O(N^{-1})$ first order term from the small-sample bias of the Maximum Likelihood Estimator. Here we show that the general Firth bias reduction technique simplifies to encouraging uniform class assignment probabilities for multinomial logistic classification, and almost has the same effect in cosine classifiers. We derive an easy-to-implement optimization objective for Firth penalized multinomial logistic and cosine classifiers, which is equivalent to penalizing the cross-entropy loss with a KL-divergence between the uniform label distribution and the predictions. Then, we empirically evaluate that it is consistently effective across the board for few-shot image classification, regardless of (1) the feature representations from different backbones, (2) the number of samples per class, and (3) the number of classes. Finally, we show the robustness of Firth bias reduction, in the case of imbalanced data distribution. Our implementation is available at https://github.com/ehsansaleh/firth_bias_reduction
翻译:在统计机器学习和计算机视觉的微缩偏差中,微缩偏差分类的性能因分类参数估计的偏差而受到影响;然而,在培训微缩分类员时,一种有效的减少偏差的基本技术却被忽略了。在这项工作中,我们展示了Firth偏差在少发分类中减少的效用。理论上,Firth偏差从统计机学习和计算机视觉的微缩偏差中消除了美元(N ⁇ -1})的第一个顺序术语,这相当于惩罚最大相似度模拟师的微缩偏差中的第一个数额($O(N ⁇ -1})。在这里,我们显示一般Firth偏差减少技术在估算分类参数时存在偏差偏差;然而,我们从直径分类中得出一个易于执行的优化目标。最后,我们从直径分类中持续地评估了我们现有精度/直径直的平流的平流率比例。最后,我们从一个直径直径分类中持续地评估了我们现有平流/直径的平流的平流比例。最后,我们从几类的平流的平流分析中,我们从直径直径的平流的平流的平流的平流的平流的平流的平流的平等的平等的平流图图图图的平流图的平流图的平流图的平流图的平流图的平流图的平流图的平比。