In recent years, we have witnessed a surge of interests in learning a suitable distance metric from weakly supervised data. Most existing methods aim to pull all the similar samples closer while push the dissimilar ones as far as possible. However, when some classes of the dataset exhibit multimodal distribution, these goals conflict and thus can hardly be concurrently satisfied. Additionally, to ensure a valid metric, many methods require a repeated eigenvalue decomposition process, which is expensive and numerically unstable. Therefore, how to learn an appropriate distance metric from weakly supervised data remains an open but challenging problem. To address this issue, in this paper, we propose a novel weakly supervised metric learning algorithm, named MultimoDal Aware weakly supervised Metric Learning (MDaML). MDaML partitions the data space into several clusters and allocates the local cluster centers and weight for each sample. Then, combining it with the weighted triplet loss can further enhance the local separability, which encourages the local dissimilar samples to keep a large distance from the local similar samples. Meanwhile, MDaML casts the metric learning problem into an unconstrained optimization on the SPD manifold, which can be efficiently solved by Riemannian Conjugate Gradient Descent (RCGD). Extensive experiments conducted on 13 datasets validate the superiority of the proposed MDaML.
翻译:近些年来,我们目睹了从监管薄弱的数据中学习适当距离指标的兴趣激增,大多数现有方法旨在拉近所有类似的样本,同时尽可能将不同样本推近。然而,当数据集的某些类别展示了多式联运分布时,这些目标发生冲突,因此很难同时实现。此外,为了确保有效的指标,许多方法需要反复采用电子价值分解过程,这个过程费用昂贵,而且数字不稳定。因此,如何从监管薄弱的数据中学习适当的距离指标仍然是一个开放但具有挑战性的问题。为了解决这一问题,我们在本文件中提出了一个新的监管不力的标准化学习算法,名为多运动了解度不强的多指标学习算法(MDAML)。MDAML将数据空间分成几个组,分配当地集束中心和每个样本的权重。然后,将它与加权三重损失结合起来,可以进一步增强本地的分解能力,从而鼓励本地的不相近的样本远离本地的类似样本。同时,MDMLML将计量学习问题引入了SDDG(G GRAMG Grod Grealm) 的不难分解的模型。