Recognizing images with long-tailed distributions remains a challenging problem while there lacks an interpretable mechanism to solve this problem. In this study, we formulate Long-tailed recognition as Domain Adaption (LDA), by modeling the long-tailed distribution as an unbalanced domain and the general distribution as a balanced domain. Within the balanced domain, we propose to slack the generalization error bound, which is defined upon the empirical risks of unbalanced and balanced domains and the divergence between them. We propose to jointly optimize empirical risks of the unbalanced and balanced domains and approximate their domain divergence by intra-class and inter-class distances, with the aim to adapt models trained on the long-tailed distribution to general distributions in an interpretable way. Experiments on benchmark datasets for image recognition, object detection, and instance segmentation validate that our LDA approach, beyond its interpretability, achieves state-of-the-art performance. Code is available at https://github.com/pengzhiliang/LDA.
翻译:在这项研究中,我们通过将长尾分配模式建模为不平衡域,将一般分布模式建模为平衡域,从而将长尾分配模式建模为平衡域;在平衡域内,我们提议放松一般化错误的界限,该界限是根据不平衡和平衡域以及域间差异的经验风险来界定的;我们提议共同优化不平衡和平衡域的经验风险,并按阶级内部和阶级间距离接近其区域差异,目的是将经过长尾分配培训的模型改编为可解释的通用分布模式;关于图像识别基准数据集的实验、对象探测和实例分割,证实我们的LDA方法除了可解释性外,还取得了最新性能;我们提议共同优化不平衡和平衡域间的经验风险,并按阶级间距离和阶级间距离接近其域间的差异,目的是将经过长尾分配模式改编成可解释的通用分布模式;关于图像识别基准数据集的实验,验证我们的LDA方法除了可解释性能外,还取得了最新性能,可在https://github.com/penghiliang/LDA上查阅。