We address the overlooked unbiasedness in existing long-tailed classification methods: we find that their overall improvement is mostly attributed to the biased preference of tail over head, as the test distribution is assumed to be balanced; however, when the test is as imbalanced as the long-tailed training data -- let the test respect Zipf's law of nature -- the tail bias is no longer beneficial overall because it hurts the head majorities. In this paper, we propose Cross-Domain Empirical Risk Minimization (xERM) for training an unbiased model to achieve strong performances on both test distributions, which empirically demonstrates that xERM fundamentally improves the classification by learning better feature representation rather than the head vs. tail game. Based on causality, we further theoretically explain why xERM achieves unbiasedness: the bias caused by the domain selection is removed by adjusting the empirical risks on the imbalanced domain and the balanced but unseen domain. Codes are available at https://github.com/BeierZhu/xERM.
翻译:我们处理现有的长尾分类方法中被忽视的公正性问题:我们发现,它们的总体改进主要归因于对尾部的偏向偏向,因为测试分布假定是均衡的;然而,当测试与长尾培训数据一样不平衡时 -- -- 让测试尊重齐普夫的自然法则 -- -- 尾部偏向在总体上不再有益,因为它伤害了多数人。在本文件中,我们提议跨尾部经验风险最小化(xEMM)来培训一种在两种测试分布上取得强力表现的不偏颇模式,这从经验上表明,XEMM通过学习更好的特征代表而不是头部对尾部游戏从根本上改进分类。基于因果关系,我们从理论上进一步解释为什么XEMM实现公正性:通过调整不平衡域和平衡但看不见域的经验风险,消除了域选择造成的偏向。代码见https://github.com/BeierZhu/xERM。