Label distribution (LD) uses the description degree to describe instances, which provides more fine-grained supervision information when learning with label ambiguity. Nevertheless, LD is unavailable in many real-world applications. To obtain LD, label enhancement (LE) has emerged to recover LD from logical label. Existing LE approach have the following problems: (\textbf{i}) They use logical label to train mappings to LD, but the supervision information is too loose, which can lead to inaccurate model prediction; (\textbf{ii}) They ignore feature redundancy and use the collected features directly. To solve (\textbf{i}), we use the topology of the feature space to generate more accurate label-confidence. To solve (\textbf{ii}), we proposed a novel supervised LE dimensionality reduction approach, which projects the original data into a lower dimensional feature space. Combining the above two, we obtain the augmented data for LE. Further, we proposed a novel nonlinear LE model based on the label-confidence and reduced features. Extensive experiments on 12 real-world datasets are conducted and the results show that our method consistently outperforms the other five comparing approaches.
翻译:标签分布(LD)使用描述度来描述实例,当学习具有标签模糊性时,它提供了更细粒度的监督信息。然而,在许多实际应用中,LD不可用。为了获得LD,标签增强(LE)已经出现,可以从逻辑标签中恢复LD。现有的LE方法存在以下问题:(\textbf{i})它们使用逻辑标签来训练映射到LD,但监督信息过于松散,可能导致模型预测不准确;(\textbf{ii})它们忽略特征冗余并直接使用收集的特征。为了解决(\textbf{i}),我们使用特征空间的拓扑来生成更准确的标签置信度。为了解决(\textbf{ii}),我们提出了一种新颖的监督LE降维方法,将原始数据投影到更低的特征空间中。将上述两种方法结合起来,我们获得了用于LE的增强数据。进一步地,我们提出了一种基于标签置信度和减少特征的新型非线性LE模型。在12个真实数据集上进行了大量实验,结果表明,我们的方法始终优于其他五种比较方法。