Deep metric learning (DML) aims to minimize empirical expected loss of the pairwise intra-/inter- class proximity violations in the embedding image. We relate DML to feasibility problem of finite chance constraints. We show that minimizer of proxy-based DML satisfies certain chance constraints, and that the worst case generalization performance of the proxy-based methods can be characterized by the radius of the smallest ball around a class proxy to cover the entire domain of the corresponding class samples, suggesting multiple proxies per class helps performance. To provide a scalable algorithm as well as exploiting more proxies, we consider the chance constraints implied by the minimizers of proxy-based DML instances and reformulate DML as finding a feasible point in intersection of such constraints, resulting in a problem to be approximately solved by iterative projections. Simply put, we repeatedly train a regularized proxy-based loss and re-initialize the proxies with the embeddings of the deliberately selected new samples. We apply our method with the well-accepted losses and evaluate on four popular benchmark datasets for image retrieval. Outperforming state-of-the-art, our method consistently improves the performance of the applied losses. Code is available at: https://github.com/yetigurbuz/ccp-dml
翻译:深度指标学习(DML)旨在最大限度地减少嵌入图像中成对类内/跨类间近距离侵犯事件的实际预期损失。我们把DML与有限机会限制的可行性问题联系起来。我们表明,以代理为基础的DML最大限度地减少某些机会限制,而以代理为基础的方法最差的概括性表现特征是,一个班级代理人周围最小球的半径覆盖相应类别样本的整个领域,建议每类多个代理人有助于业绩。为了提供一种可扩缩的算法以及利用更多的代理人,我们认为,以代理为基础的DML事件最小化者所隐含的机会限制以及将DML改造成的机会限制作为在这种制约因素交汇处找到一个可行的点,从而导致一个问题可以通过迭接预测大致解决。简而言之,我们反复培训一种正规化的以代理为基础的损失,并以有意选取的新样品的嵌入方式重新开启轴线。我们采用的方法对公认的损失进行评估,并评估四种受欢迎的基准数据集进行图像检索。超状态-艺术,我们的方法持续改进了MAGI/Cmbrm 。