In person re-identification (re-ID) task, it is still challenging to learn discriminative representation by deep learning, due to limited data. Generally speaking, the model will get better performance when increasing the amount of data. The addition of similar classes strengthens the ability of the classifier to identify similar identities, thereby improving the discrimination of representation. In this paper, we propose a Diverse and Compact Transformer (DC-Former) that can achieve a similar effect by splitting embedding space into multiple diverse and compact subspaces. Compact embedding subspace helps model learn more robust and discriminative embedding to identify similar classes. And the fusion of these diverse embeddings containing more fine-grained information can further improve the effect of re-ID. Specifically, multiple class tokens are used in vision transformer to represent multiple embedding spaces. Then, a self-diverse constraint (SDC) is applied to these spaces to push them away from each other, which makes each embedding space diverse and compact. Further, a dynamic weight controller(DWC) is further designed for balancing the relative importance among them during training. The experimental results of our method are promising, which surpass previous state-of-the-art methods on several commonly used person re-ID benchmarks.
翻译:个人再身份( re-ID) 任务中,由于数据有限,通过深度学习来学习歧视性代表,仍然具有挑战性。一般来说,模型在增加数据数量时会取得更好的性能。增加类似类别会加强分类者识别类似身份的能力,从而改善代表性的区别。在本文件中,我们提议了多样化和契约变换器(DC-Former),通过将空间分割成多种不同和紧凑的子空间,可以实现类似效果。契约嵌入子空间有助于模型学习更强大和有区别的嵌入,以识别相似的类别。而包含更精细的信息的这些多样化嵌入器的结合可以进一步改善再识别数据的效果。具体地说,在视觉变压器中使用多个类符号来代表多个嵌入空间。然后,对这些空间应用了自我反向约束(SDC),以将它们从彼此分离出去,使每个嵌入的空间多样化和紧凑。此外,动态重重力控制器(DWC)正在进一步设计,以平衡它们在培训期间的相对重要性。我们所采用的方法的实验性结果比以往更具有前景。</s>