Recently, vehicle re-identification methods based on deep learning constitute remarkable achievement. However, this achievement requires large-scale and well-annotated datasets. In constructing the dataset, assigning globally available identities (Ids) to vehicles captured from a great number of cameras is labour-intensive, because it needs to consider their subtle appearance differences or viewpoint variations. In this paper, we propose camera-tracklet-aware contrastive learning (CTACL) using the multi-camera tracklet information without vehicle identity labels. The proposed CTACL divides an unlabelled domain, i.e., entire vehicle images, into multiple camera-level subdomains and conducts contrastive learning within and beyond the subdomains. The positive and negative samples for contrastive learning are defined using tracklet Ids of each camera. Additionally, the domain adaptation across camera networks is introduced to improve the generalisation performance of learnt representations and alleviate the performance degradation resulted from the domain gap between the subdomains. We demonstrate the effectiveness of our approach on video-based and image-based vehicle Re-ID datasets. Experimental results show that the proposed method outperforms the recent state-of-the-art unsupervised vehicle Re-ID methods. The source code for this paper is publicly available on `https://github.com/andreYoo/CTAM-CTACL-VVReID.git'.
翻译:最近,基于深层学习的车辆再识别方法是令人瞩目的成就,然而,这一成就需要大规模和附加说明的数据集。在建立数据集时,为从大量照相机中捕获的车辆分配全球可用身份(IDs)是劳动密集型的,因为它需要考虑其微妙的外观差异或观点差异。在本文中,我们提议使用没有车辆身份标签的多摄像轨道信息进行相机-轨道对比学习(CTACL)。拟议的CTOCL将一个未标记的域(即整个车辆图像)分为多个摄像头级子域,并在子域内外进行对比性学习。对比性学习的正面和负面样本是使用每台相机的轨道Ids定义的。此外,我们采用跨相机网络的域调整来改进所学演示的通用性表现,并减轻子域隔隔隔的性性差造成的性能退化。我们展示了我们对基于视频的和基于图像的车辆重新ID数据集的处理方法的有效性。实验结果显示,用于对比性学习的样本学习的样本样本样本/Regibr