We propose GradTail, an algorithm that uses gradients to improve model performance on the fly in the face of long-tailed training data distributions. Unlike conventional long-tail classifiers which operate on converged - and possibly overfit - models, we demonstrate that an approach based on gradient dot product agreement can isolate long-tailed data early on during model training and improve performance by dynamically picking higher sample weights for that data. We show that such upweighting leads to model improvements for both classification and regression models, the latter of which are relatively unexplored in the long-tail literature, and that the long-tail examples found by gradient alignment are consistent with our semantic expectations.
翻译:我们建议GradTail(GradTail ), 这一算法使用梯度来改善长期培训数据分布条件下的飞行模型性能。 与传统的长尾分类方法不同,这些分类方法以趋同(甚至超合适)模式运作,我们证明基于梯度点产品协议的方法可以在模型培训初期就将长尾数据分离出来,并通过动态地为这些数据采集更高的样本权重来改进性能。 我们发现,这种加权方法导致分类和回归模型的模型改进,后者在长尾文献中相对没有被探讨,而梯度调整发现的长期性实例符合我们的语义期望。