Complementary-label learning (CLL) is a common application in the scenario of weak supervision. However, in real-world datasets, CLL encounters class-imbalanced training samples, where the quantity of samples of one class is significantly lower than those of other classes. Unfortunately, existing CLL approaches have yet to explore the problem of class-imbalanced samples, which reduces the prediction accuracy, especially in imbalanced classes. In this paper, we propose a novel problem setting to allow learning from class-imbalanced complementarily labeled samples for multi-class classification. Accordingly, to deal with this novel problem, we propose a new CLL approach, called Weighted Complementary-Label Learning (WCLL). The proposed method models a weighted empirical risk minimization loss by utilizing the class-imbalanced complementarily labeled information, which is also applicable to multi-class imbalanced training samples. Furthermore, the estimation error bound of the proposed method was derived to provide a theoretical guarantee. Finally, we do extensive experiments on widely-used benchmark datasets to validate the superiority of our method by comparing it with existing state-of-the-art methods.
翻译:补充标签学习(CLL)是监督不力情况下的一种常见应用。然而,在现实世界的数据集中,CLL会遇到类平衡培训样本,其中某一类样本的数量大大低于其他类样本的数量。不幸的是,现有的CLL办法尚未探讨类平衡抽样问题,这降低了预测的准确性,特别是在不平衡的类别中。在本文中,我们提出了一个新问题设置,以便从类平衡的贴标签样本中学习多类分类。因此,为了处理这个新的问题,我们提出了新的CLL(CLL)办法,称为Weight Cuply-Label Learning(WLL) 。拟议方法模型采用加权经验风险最小化模型,使用类平衡的补充标签信息,这也适用于多类不平衡培训样本。此外,拟议方法的估算误差是为了提供理论保证。最后,我们对广泛使用的基准数据集进行了广泛的实验,以证实我们方法的优越性,将之与现有先进方法进行比较。