Person images captured by surveillance cameras are often occluded by various obstacles, which lead to defective feature representation and harm person re-identification (Re-ID) performance. To tackle this challenge, we propose to reconstruct the feature representation of occluded parts by fully exploiting the information of its neighborhood in a gallery image set. Specifically, we first introduce a visible part-based feature by body mask for each person image. Then we identify its neighboring samples using the visible features and reconstruct the representation of the full body by an outlier-removable graph neural network with all the neighboring samples as input. Extensive experiments show that the proposed approach obtains significant improvements. In the large-scale Occluded-DukeMTMC benchmark, our approach achieves 64.2% mAP and 67.6% rank-1 accuracy which outperforms the state-of-the-art approaches by large margins, i.e.,20.4% and 12.5%, respectively, indicating the effectiveness of our method on occluded Re-ID problem.
翻译:由监视摄像机拍摄的个人图像往往被各种障碍所掩盖,这些障碍导致特征描述有缺陷,伤害个人重新识别(Re-ID)性能。为了应对这一挑战,我们提议通过在画廊图像集中充分利用其周边信息来重建隐蔽部分的特征描述。具体地说,我们首先为每个人的图像引入一个以身体遮罩为主的可见部分特征。然后,我们利用这些可见特征来识别其周边样本,并用一个外向可移动的图形神经网络来重建整个身体的特征描述,将所有相邻样本作为输入。广泛的实验表明,拟议的方法取得了显著的改进。在大型的Occclone-DukeMTMC基准中,我们的方法达到了64.2% mAP和67.6%-1级的精度,这些精度分别比大型边缘(即20.4%和12.5%)的先进方法更优于最先进的方法,表明我们应对隐蔽的重新识别问题的方法的有效性。