Vehicle re-identification (reID) often requires recognize a target vehicle in large datasets captured from multi-cameras. It plays an important role in the automatic analysis of the increasing urban surveillance videos, which has become a hot topic in recent years. However, the appearance of vehicle images is easily affected by the environment that various illuminations, different backgrounds and viewpoints, which leads to the large bias between different cameras. To address this problem, this paper proposes a cross-camera adaptation framework (CCA), which smooths the bias by exploiting the common space between cameras for all samples. CCA first transfers images from multi-cameras into one camera to reduce the impact of the illumination and resolution, which generates the samples with the similar distribution. Then, to eliminate the influence of background and focus on the valuable parts, we propose an attention alignment network (AANet) to learn powerful features for vehicle reID. Specially, in AANet, the spatial transfer network with attention module is introduced to locate a series of the most discriminative regions with high-attention weights and suppress the background. Moreover, comprehensive experimental results have demonstrated that our proposed CCA can achieve excellent performances on benchmark datasets VehicleID and VeRi-776.
翻译:车辆再识别(reID)往往要求在从多镜头采集的大型数据集中识别目标车辆,在自动分析不断增长的城市监视录像中发挥重要作用,这种录像近年来已成为热题,但车辆图像的外观很容易受到各种照明、不同背景和观点导致不同相机之间产生巨大偏差的环境的影响。为解决这一问题,本文件提议了一个跨镜头适应框架,通过利用所有样品的照相机之间的共同空间来消除偏差。共同国家评估首先将多摄像头的图像传送到一个照相机中,以减少照明和分辨率的影响,从而产生类似分布的样品。然后,为了消除背景影响和对有价值的部分的关注,我们提议建立一个关注调整网络(AANet),以学习车辆再识别的强大特征。特别是,在AANet中,引入了带有关注模块的空间传输网络,以找到一系列具有高留量和抑制背景的最有歧视的区域。此外,全面的实验结果表明,我们拟议的CAID76和VeR飞行器可以在基准数据设置上取得出色的业绩。