Triplet learning, i.e. learning from triplet data, has attracted much attention in computer vision tasks with an extremely large number of categories, e.g., face recognition and person re-identification. Albeit with rapid progress in designing and applying triplet learning algorithms, there is a lacking study on the theoretical understanding of their generalization performance. To fill this gap, this paper investigates the generalization guarantees of triplet learning by leveraging the stability analysis. Specifically, we establish the first general high-probability generalization bound for the triplet learning algorithm satisfying the uniform stability, and then obtain the excess risk bounds of the order $O(n^{-\frac{1}{2}} \mathrm{log}n)$ for both stochastic gradient descent (SGD) and regularized risk minimization (RRM), where $2n$ is approximately equal to the number of training samples. Moreover, an optimistic generalization bound in expectation as fast as $O(n^{-1})$ is derived for RRM in a low noise case via the on-average stability analysis. Finally, our results are applied to triplet metric learning to characterize its theoretical underpinning.
翻译:三联学习,即从三联数据中学习,在计算机视野任务中引起了极大关注,其类别非常多,例如面部识别和个人再识别。尽管在设计和应用三联学习算法方面进展迅速,但对其一般化表现的理论理解却缺乏研究。为填补这一空白,本文件调查了利用稳定性分析来普遍三联学习的保障。具体地说,我们为符合统一稳定性的三联学习算法建立了第一个通用的通用高概率通用,然后获得美元(n ⁇ -\frac{1 ⁇ 2 ⁇ \mathrm{log}n)的超风险界限。最后,我们的结果应用于三联梯梯度梯度脱落和常规化风险最小化,其中2n美元大约相当于培训样本的数量。此外,在低噪音案件中,我们的结果被应用到三联度标准,用于理论基础。