In this work we present a novel unsupervised framework for hard training example mining. The only input to the method is a collection of images relevant to the target application and a meaningful initial representation, provided e.g. by pre-trained CNN. Positive examples are distant points on a single manifold, while negative examples are nearby points on different manifolds. Both types of examples are revealed by disagreements between Euclidean and manifold similarities. The discovered examples can be used in training with any discriminative loss. The method is applied to unsupervised fine-tuning of pre-trained networks for fine-grained classification and particular object retrieval. Our models are on par or are outperforming prior models that are fully or partially supervised.
翻译:在这项工作中,我们为硬训练采矿提供了一个新的、不受监督的框架。对方法的唯一投入是收集与目标应用有关的图像和有意义的初步表述,例如由经过预先训练的CNN提供。正面的例子是单方形的遥远点,而负面的例子则是不同方形的近点。两种例子都通过欧几里得和多重相似性之间的分歧而揭示。发现的例子可以用于任何歧视性损失的培训中。这种方法用于未经监督的精细分类和特定对象检索培训前网络的微调。我们的模型是完全或部分监督的,或优于以往的模型。